blob: fb38b5f267b2d4185dddd475c42d56e9ebf2efd7 [file] [log] [blame]
/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.cassandra.db;
import java.io.DataInput;
import java.io.DataOutput;
import java.io.IOException;
import java.nio.ByteBuffer;
import org.apache.cassandra.io.ISerializer;
import org.apache.cassandra.io.FSReadError;
import org.apache.cassandra.io.util.FileDataInput;
import org.apache.cassandra.utils.ByteBufferUtil;
public class ColumnSerializer implements ISerializer<Column>
{
public final static int DELETION_MASK = 0x01;
public final static int EXPIRATION_MASK = 0x02;
public final static int COUNTER_MASK = 0x04;
public final static int COUNTER_UPDATE_MASK = 0x08;
public final static int RANGE_TOMBSTONE_MASK = 0x10;
/**
* Flag affecting deserialization behavior.
* - LOCAL: for deserialization of local data (Expired columns are
* converted to tombstones (to gain disk space)).
* - FROM_REMOTE: for deserialization of data received from remote hosts
* (Expired columns are converted to tombstone and counters have
* their delta cleared)
* - PRESERVE_SIZE: used when no transformation must be performed, i.e,
* when we must ensure that deserializing and reserializing the
* result yield the exact same bytes. Streaming uses this.
*/
public static enum Flag
{
LOCAL, FROM_REMOTE, PRESERVE_SIZE;
}
public void serialize(Column column, DataOutput out) throws IOException
{
assert column.name().remaining() > 0;
ByteBufferUtil.writeWithShortLength(column.name(), out);
try
{
out.writeByte(column.serializationFlags());
if (column instanceof CounterColumn)
{
out.writeLong(((CounterColumn)column).timestampOfLastDelete());
}
else if (column instanceof ExpiringColumn)
{
out.writeInt(((ExpiringColumn) column).getTimeToLive());
out.writeInt(column.getLocalDeletionTime());
}
out.writeLong(column.timestamp());
ByteBufferUtil.writeWithLength(column.value(), out);
}
catch (IOException e)
{
throw new RuntimeException(e);
}
}
public Column deserialize(DataInput in) throws IOException
{
return deserialize(in, Flag.LOCAL);
}
/*
* For counter columns, we must know when we deserialize them if what we
* deserialize comes from a remote host. If it does, then we must clear
* the delta.
*/
public Column deserialize(DataInput in, ColumnSerializer.Flag flag) throws IOException
{
return deserialize(in, flag, Integer.MIN_VALUE);
}
public Column deserialize(DataInput in, ColumnSerializer.Flag flag, int expireBefore) throws IOException
{
ByteBuffer name = ByteBufferUtil.readWithShortLength(in);
if (name.remaining() <= 0)
throw CorruptColumnException.create(in, name);
int b = in.readUnsignedByte();
return deserializeColumnBody(in, name, b, flag, expireBefore);
}
Column deserializeColumnBody(DataInput in, ByteBuffer name, int mask, ColumnSerializer.Flag flag, int expireBefore) throws IOException
{
if ((mask & COUNTER_MASK) != 0)
{
long timestampOfLastDelete = in.readLong();
long ts = in.readLong();
ByteBuffer value = ByteBufferUtil.readWithLength(in);
return CounterColumn.create(name, value, ts, timestampOfLastDelete, flag);
}
else if ((mask & EXPIRATION_MASK) != 0)
{
int ttl = in.readInt();
int expiration = in.readInt();
long ts = in.readLong();
ByteBuffer value = ByteBufferUtil.readWithLength(in);
return ExpiringColumn.create(name, value, ts, ttl, expiration, expireBefore, flag);
}
else
{
long ts = in.readLong();
ByteBuffer value = ByteBufferUtil.readWithLength(in);
return (mask & COUNTER_UPDATE_MASK) != 0
? new CounterUpdateColumn(name, value, ts)
: ((mask & DELETION_MASK) == 0
? new Column(name, value, ts)
: new DeletedColumn(name, value, ts));
}
}
public long serializedSize(Column column, TypeSizes type)
{
return column.serializedSize(type);
}
public static class CorruptColumnException extends IOException
{
public CorruptColumnException(String s)
{
super(s);
}
public static CorruptColumnException create(DataInput in, ByteBuffer name)
{
assert name.remaining() <= 0;
String format = "invalid column name length %d%s";
String details = "";
if (in instanceof FileDataInput)
{
FileDataInput fdis = (FileDataInput)in;
long remaining;
try
{
remaining = fdis.bytesRemaining();
}
catch (IOException e)
{
throw new FSReadError(e, fdis.getPath());
}
details = String.format(" (%s, %d bytes remaining)", fdis.getPath(), remaining);
}
return new CorruptColumnException(String.format(format, name.remaining(), details));
}
}
}