blob: c242f010050512648d1d86eb41f80ca0d1f1d688 [file] [log] [blame]
/**
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.gora.cassandra.store;
import java.io.IOException;
import java.nio.ByteBuffer;
import java.util.ArrayList;
import java.util.Iterator;
import java.util.LinkedHashMap;
import java.util.List;
import java.util.Map;
import java.util.Properties;
import java.util.Set;
import me.prettyprint.cassandra.serializers.IntegerSerializer;
import me.prettyprint.cassandra.serializers.StringSerializer;
import me.prettyprint.hector.api.beans.ColumnSlice;
import me.prettyprint.hector.api.beans.HColumn;
import me.prettyprint.hector.api.beans.HSuperColumn;
import me.prettyprint.hector.api.beans.Row;
import me.prettyprint.hector.api.beans.SuperRow;
import me.prettyprint.hector.api.beans.SuperSlice;
import org.apache.avro.Schema;
import org.apache.avro.Schema.Field;
import org.apache.avro.Schema.Type;
import org.apache.avro.generic.GenericArray;
import org.apache.avro.specific.SpecificFixed;
import org.apache.avro.util.Utf8;
import org.apache.gora.cassandra.query.CassandraQuery;
import org.apache.gora.cassandra.query.CassandraResult;
import org.apache.gora.cassandra.query.CassandraResultSet;
import org.apache.gora.cassandra.query.CassandraRow;
import org.apache.gora.cassandra.query.CassandraSubColumn;
import org.apache.gora.cassandra.query.CassandraSuperColumn;
import org.apache.gora.persistency.ListGenericArray;
import org.apache.gora.persistency.Persistent;
import org.apache.gora.persistency.StatefulHashMap;
import org.apache.gora.persistency.impl.PersistentBase;
import org.apache.gora.persistency.impl.StateManagerImpl;
import org.apache.gora.query.PartitionQuery;
import org.apache.gora.query.Query;
import org.apache.gora.query.Result;
import org.apache.gora.query.impl.PartitionQueryImpl;
import org.apache.gora.store.impl.DataStoreBase;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;
public class CassandraStore<K, T extends Persistent> extends DataStoreBase<K, T> {
public static final Logger LOG = LoggerFactory.getLogger(CassandraStore.class);
private CassandraClient<K, T> cassandraClient = new CassandraClient<K, T>();
/**
* The values are Avro fields pending to be stored.
*
* We want to iterate over the keys in insertion order.
* We don't want to lock the entire collection before iterating over the keys, since in the meantime other threads are adding entries to the map.
*/
private Map<K, T> buffer = new LinkedHashMap<K, T>();
public CassandraStore() throws Exception {
// this.cassandraClient.initialize();
}
public void initialize(Class<K> keyClass, Class<T> persistent, Properties properties) throws IOException {
super.initialize(keyClass, persistent, properties);
try {
this.cassandraClient.initialize(keyClass, persistent);
}
catch (Exception e) {
throw new IOException(e.getMessage(), e);
}
}
@Override
public void close() throws IOException {
LOG.debug("close");
flush();
}
@Override
public void createSchema() {
LOG.debug("creating Cassandra keyspace");
this.cassandraClient.checkKeyspace();
}
@Override
public boolean delete(K key) throws IOException {
LOG.debug("delete " + key);
return false;
}
@Override
public long deleteByQuery(Query<K, T> query) throws IOException {
LOG.debug("delete by query " + query);
return 0;
}
@Override
public void deleteSchema() throws IOException {
LOG.debug("delete schema");
this.cassandraClient.dropKeyspace();
}
@Override
public Result<K, T> execute(Query<K, T> query) throws IOException {
Map<String, List<String>> familyMap = this.cassandraClient.getFamilyMap(query);
Map<String, String> reverseMap = this.cassandraClient.getReverseMap(query);
CassandraQuery<K, T> cassandraQuery = new CassandraQuery<K, T>();
cassandraQuery.setQuery(query);
cassandraQuery.setFamilyMap(familyMap);
CassandraResult<K, T> cassandraResult = new CassandraResult<K, T>(this, query);
cassandraResult.setReverseMap(reverseMap);
CassandraResultSet cassandraResultSet = new CassandraResultSet();
// We query Cassandra keyspace by families.
for (String family : familyMap.keySet()) {
if (family == null) {
continue;
}
if (this.cassandraClient.isSuper(family)) {
addSuperColumns(family, cassandraQuery, cassandraResultSet);
} else {
addSubColumns(family, cassandraQuery, cassandraResultSet);
}
}
cassandraResult.setResultSet(cassandraResultSet);
return cassandraResult;
}
private void addSubColumns(String family, CassandraQuery<K, T> cassandraQuery,
CassandraResultSet cassandraResultSet) {
// select family columns that are included in the query
List<Row<K, ByteBuffer, ByteBuffer>> rows = this.cassandraClient.execute(cassandraQuery, family);
for (Row<K, ByteBuffer, ByteBuffer> row : rows) {
K key = row.getKey();
// find associated row in the resultset
CassandraRow<K> cassandraRow = cassandraResultSet.getRow(key);
if (cassandraRow == null) {
cassandraRow = new CassandraRow<K>();
cassandraResultSet.putRow(key, cassandraRow);
cassandraRow.setKey(key);
}
ColumnSlice<ByteBuffer, ByteBuffer> columnSlice = row.getColumnSlice();
for (HColumn<ByteBuffer, ByteBuffer> hColumn : columnSlice.getColumns()) {
CassandraSubColumn cassandraSubColumn = new CassandraSubColumn();
cassandraSubColumn.setValue(hColumn);
cassandraSubColumn.setFamily(family);
cassandraRow.add(cassandraSubColumn);
}
}
}
private void addSuperColumns(String family, CassandraQuery<K, T> cassandraQuery,
CassandraResultSet cassandraResultSet) {
List<SuperRow<K, String, ByteBuffer, ByteBuffer>> superRows = this.cassandraClient.executeSuper(cassandraQuery, family);
for (SuperRow<K, String, ByteBuffer, ByteBuffer> superRow: superRows) {
K key = superRow.getKey();
CassandraRow<K> cassandraRow = cassandraResultSet.getRow(key);
if (cassandraRow == null) {
cassandraRow = new CassandraRow();
cassandraResultSet.putRow(key, cassandraRow);
cassandraRow.setKey(key);
}
SuperSlice<String, ByteBuffer, ByteBuffer> superSlice = superRow.getSuperSlice();
for (HSuperColumn<String, ByteBuffer, ByteBuffer> hSuperColumn: superSlice.getSuperColumns()) {
CassandraSuperColumn cassandraSuperColumn = new CassandraSuperColumn();
cassandraSuperColumn.setValue(hSuperColumn);
cassandraSuperColumn.setFamily(family);
cassandraRow.add(cassandraSuperColumn);
}
}
}
/**
* Flush the buffer. Write the buffered rows.
* @see org.apache.gora.store.DataStore#flush()
*/
@Override
public void flush() throws IOException {
Set<K> keys = this.buffer.keySet();
// this duplicates memory footprint
K[] keyArray = (K[]) keys.toArray();
// iterating over the key set directly would throw ConcurrentModificationException with java.util.HashMap and subclasses
for (K key: keyArray) {
T value = this.buffer.get(key);
if (value == null) {
LOG.info("Value to update is null for key " + key);
continue;
}
Schema schema = value.getSchema();
for (Field field: schema.getFields()) {
if (value.isDirty(field.pos())) {
addOrUpdateField(key, field, value.get(field.pos()));
}
}
}
// remove flushed rows
for (K key: keyArray) {
this.buffer.remove(key);
}
}
@Override
public T get(K key, String[] fields) throws IOException {
CassandraQuery<K,T> query = new CassandraQuery<K,T>();
query.setDataStore(this);
query.setKeyRange(key, key);
query.setFields(fields);
query.setLimit(1);
Result<K,T> result = execute(query);
boolean hasResult = result.next();
return hasResult ? result.get() : null;
}
@Override
public List<PartitionQuery<K, T>> getPartitions(Query<K, T> query)
throws IOException {
// just a single partition
List<PartitionQuery<K,T>> partitions = new ArrayList<PartitionQuery<K,T>>();
partitions.add(new PartitionQueryImpl<K,T>(query));
return partitions;
}
/**
* In Cassandra Schemas are referred to as Keyspaces
* @return Keyspace
*/
@Override
public String getSchemaName() {
return this.cassandraClient.getKeyspaceName();
}
@Override
public Query<K, T> newQuery() {
Query<K,T> query = new CassandraQuery<K, T>(this);
query.setFields(getFieldsToQuery(null));
return query;
}
/**
* Duplicate instance to keep all the objects in memory till flushing.
* @see org.apache.gora.store.DataStore#put(java.lang.Object, org.apache.gora.persistency.Persistent)
*/
@Override
public void put(K key, T value) throws IOException {
T p = (T) value.newInstance(new StateManagerImpl());
Schema schema = value.getSchema();
for (Field field: schema.getFields()) {
int fieldPos = field.pos();
if (value.isDirty(fieldPos)) {
Object fieldValue = value.get(fieldPos);
// check if field has a nested structure (array, map, or record)
Schema fieldSchema = field.schema();
Type type = fieldSchema.getType();
switch(type) {
case RECORD:
Persistent persistent = (Persistent) fieldValue;
Persistent newRecord = persistent.newInstance(new StateManagerImpl());
for (Field member: fieldSchema.getFields()) {
newRecord.put(member.pos(), persistent.get(member.pos()));
}
fieldValue = newRecord;
break;
case MAP:
// needs to keep State.DELETED.
break;
case ARRAY:
GenericArray array = (GenericArray) fieldValue;
ListGenericArray newArray = new ListGenericArray(fieldSchema.getElementType());
Iterator iter = array.iterator();
while (iter.hasNext()) {
newArray.add(iter.next());
}
fieldValue = newArray;
break;
}
p.put(fieldPos, fieldValue);
}
}
// this performs a structural modification of the map
this.buffer.put(key, p);
}
/**
* Add a field to Cassandra according to its type.
* @param key the key of the row where the field should be added
* @param field the Avro field representing a datum
* @param value the field value
*/
private void addOrUpdateField(K key, Field field, Object value) {
Schema schema = field.schema();
Type type = schema.getType();
switch (type) {
case STRING:
case BOOLEAN:
case INT:
case LONG:
case BYTES:
case FLOAT:
case DOUBLE:
case FIXED:
this.cassandraClient.addColumn(key, field.name(), value);
break;
case RECORD:
if (value != null) {
if (value instanceof PersistentBase) {
PersistentBase persistentBase = (PersistentBase) value;
for (Field member: schema.getFields()) {
// TODO: hack, do not store empty arrays
Object memberValue = persistentBase.get(member.pos());
if (memberValue instanceof GenericArray<?>) {
if (((GenericArray)memberValue).size() == 0) {
continue;
}
} else if (memberValue instanceof StatefulHashMap<?,?>) {
if (((StatefulHashMap)memberValue).size() == 0) {
continue;
}
}
this.cassandraClient.addSubColumn(key, field.name(), member.name(), memberValue);
}
} else {
LOG.info("Record not supported: " + value.toString());
}
}
break;
case MAP:
if (value != null) {
if (value instanceof StatefulHashMap<?, ?>) {
this.cassandraClient.addStatefulHashMap(key, field.name(), (StatefulHashMap<Utf8,Object>)value);
} else {
LOG.info("Map not supported: " + value.toString());
}
}
break;
case ARRAY:
if (value != null) {
if (value instanceof GenericArray<?>) {
this.cassandraClient.addGenericArray(key, field.name(), (GenericArray)value);
} else {
LOG.info("Array not supported: " + value.toString());
}
}
break;
default:
LOG.info("Type not considered: " + type.name());
}
}
@Override
public boolean schemaExists() throws IOException {
LOG.info("schema exists");
return cassandraClient.keyspaceExists();
}
}