blob: 4164bce6afd8dd1680d79f219318d0a68e930d1c [file] [log] [blame]
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.flink.streaming.connectors.cassandra;
import org.apache.flink.api.java.tuple.Tuple;
/**
* Sink to write Flink {@link Tuple}s into a Cassandra cluster.
*
* @param <IN> Type of the elements emitted by this sink, it must extend {@link Tuple}
*/
public class CassandraTupleSink<IN extends Tuple> extends AbstractCassandraTupleSink<IN> {
public CassandraTupleSink(
String insertQuery,
ClusterBuilder builder) {
this(insertQuery, builder, CassandraSinkBaseConfig.newBuilder().build());
}
CassandraTupleSink(
String insertQuery,
ClusterBuilder builder,
CassandraSinkBaseConfig config) {
this(insertQuery, builder, config, new NoOpCassandraFailureHandler());
}
CassandraTupleSink(
String insertQuery,
ClusterBuilder builder,
CassandraSinkBaseConfig config,
CassandraFailureHandler failureHandler) {
super(insertQuery, builder, config, failureHandler);
}
@Override
protected Object[] extract(IN record) {
Object[] al = new Object[record.getArity()];
for (int i = 0; i < record.getArity(); i++) {
al[i] = record.getField(i);
}
return al;
}
}