| /* |
| * Licensed to the Apache Software Foundation (ASF) under one or more |
| * contributor license agreements. See the NOTICE file distributed with |
| * this work for additional information regarding copyright ownership. |
| * The ASF licenses this file to You under the Apache License, Version 2.0 |
| * (the "License"); you may not use this file except in compliance with |
| * the License. You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| package org.apache.flink.streaming.connectors.kafka; |
| |
| import org.apache.flink.api.common.serialization.DeserializationSchema; |
| import org.apache.flink.networking.NetworkFailuresProxy; |
| import org.apache.flink.streaming.api.datastream.DataStream; |
| import org.apache.flink.streaming.api.datastream.DataStreamSink; |
| import org.apache.flink.streaming.api.operators.StreamSink; |
| import org.apache.flink.streaming.connectors.kafka.partitioner.FlinkKafkaPartitioner; |
| import org.apache.flink.streaming.util.serialization.KeyedDeserializationSchema; |
| import org.apache.flink.streaming.util.serialization.KeyedDeserializationSchemaWrapper; |
| import org.apache.flink.streaming.util.serialization.KeyedSerializationSchema; |
| |
| import kafka.server.KafkaServer; |
| import org.apache.kafka.clients.consumer.ConsumerRecord; |
| |
| import java.util.ArrayList; |
| import java.util.Collection; |
| import java.util.Collections; |
| import java.util.List; |
| import java.util.Properties; |
| |
| /** |
| * Abstract class providing a Kafka test environment. |
| */ |
| public abstract class KafkaTestEnvironment { |
| /** |
| * Configuration class for {@link KafkaTestEnvironment}. |
| */ |
| public static class Config { |
| private int kafkaServersNumber = 1; |
| private Properties kafkaServerProperties = null; |
| private boolean secureMode = false; |
| private boolean hideKafkaBehindProxy = false; |
| |
| /** |
| * Please use {@link KafkaTestEnvironment#createConfig()} method. |
| */ |
| private Config() { |
| } |
| |
| public int getKafkaServersNumber() { |
| return kafkaServersNumber; |
| } |
| |
| public Config setKafkaServersNumber(int kafkaServersNumber) { |
| this.kafkaServersNumber = kafkaServersNumber; |
| return this; |
| } |
| |
| public Properties getKafkaServerProperties() { |
| return kafkaServerProperties; |
| } |
| |
| public Config setKafkaServerProperties(Properties kafkaServerProperties) { |
| this.kafkaServerProperties = kafkaServerProperties; |
| return this; |
| } |
| |
| public boolean isSecureMode() { |
| return secureMode; |
| } |
| |
| public Config setSecureMode(boolean secureMode) { |
| this.secureMode = secureMode; |
| return this; |
| } |
| |
| public boolean isHideKafkaBehindProxy() { |
| return hideKafkaBehindProxy; |
| } |
| |
| public Config setHideKafkaBehindProxy(boolean hideKafkaBehindProxy) { |
| this.hideKafkaBehindProxy = hideKafkaBehindProxy; |
| return this; |
| } |
| } |
| |
| protected static final String KAFKA_HOST = "localhost"; |
| |
| protected final List<NetworkFailuresProxy> networkFailuresProxies = new ArrayList<>(); |
| |
| public static Config createConfig() { |
| return new Config(); |
| } |
| |
| public abstract void prepare(Config config); |
| |
| public void shutdown() throws Exception { |
| for (NetworkFailuresProxy proxy : networkFailuresProxies) { |
| proxy.close(); |
| } |
| } |
| |
| public abstract void deleteTestTopic(String topic); |
| |
| public abstract void createTestTopic(String topic, int numberOfPartitions, int replicationFactor, Properties properties); |
| |
| public void createTestTopic(String topic, int numberOfPartitions, int replicationFactor) { |
| this.createTestTopic(topic, numberOfPartitions, replicationFactor, new Properties()); |
| } |
| |
| public abstract Properties getStandardProperties(); |
| |
| public abstract Properties getSecureProperties(); |
| |
| public abstract String getBrokerConnectionString(); |
| |
| public abstract String getVersion(); |
| |
| public abstract List<KafkaServer> getBrokers(); |
| |
| // -- consumer / producer instances: |
| public <T> FlinkKafkaConsumerBase<T> getConsumer(List<String> topics, DeserializationSchema<T> deserializationSchema, Properties props) { |
| return getConsumer(topics, new KeyedDeserializationSchemaWrapper<T>(deserializationSchema), props); |
| } |
| |
| public <T> FlinkKafkaConsumerBase<T> getConsumer(String topic, KeyedDeserializationSchema<T> readSchema, Properties props) { |
| return getConsumer(Collections.singletonList(topic), readSchema, props); |
| } |
| |
| public <T> FlinkKafkaConsumerBase<T> getConsumer(String topic, DeserializationSchema<T> deserializationSchema, Properties props) { |
| return getConsumer(Collections.singletonList(topic), deserializationSchema, props); |
| } |
| |
| public abstract <T> FlinkKafkaConsumerBase<T> getConsumer(List<String> topics, KeyedDeserializationSchema<T> readSchema, Properties props); |
| |
| public abstract <K, V> Collection<ConsumerRecord<K, V>> getAllRecordsFromTopic( |
| Properties properties, |
| String topic, |
| int partition, |
| long timeout); |
| |
| public abstract <T> StreamSink<T> getProducerSink(String topic, |
| KeyedSerializationSchema<T> serSchema, Properties props, |
| FlinkKafkaPartitioner<T> partitioner); |
| |
| public abstract <T> DataStreamSink<T> produceIntoKafka(DataStream<T> stream, String topic, |
| KeyedSerializationSchema<T> serSchema, Properties props, |
| FlinkKafkaPartitioner<T> partitioner); |
| |
| public abstract <T> DataStreamSink<T> writeToKafkaWithTimestamps( |
| DataStream<T> stream, |
| String topic, |
| KeyedSerializationSchema<T> serSchema, |
| Properties props); |
| |
| // -- offset handlers |
| |
| /** |
| * Simple interface to commit and retrieve offsets. |
| */ |
| public interface KafkaOffsetHandler { |
| Long getCommittedOffset(String topicName, int partition); |
| |
| void setCommittedOffset(String topicName, int partition, long offset); |
| |
| void close(); |
| } |
| |
| public abstract KafkaOffsetHandler createOffsetHandler(); |
| |
| // -- leader failure simulation |
| |
| public abstract void restartBroker(int leaderId) throws Exception; |
| |
| public abstract int getLeaderToShutDown(String topic) throws Exception; |
| |
| public abstract int getBrokerId(KafkaServer server); |
| |
| public abstract boolean isSecureRunSupported(); |
| |
| public void blockProxyTraffic() { |
| for (NetworkFailuresProxy proxy : networkFailuresProxies) { |
| proxy.blockTraffic(); |
| } |
| } |
| |
| public void unblockProxyTraffic() { |
| for (NetworkFailuresProxy proxy : networkFailuresProxies) { |
| proxy.unblockTraffic(); |
| } |
| } |
| |
| protected NetworkFailuresProxy createProxy(String remoteHost, int remotePort) { |
| NetworkFailuresProxy proxy = new NetworkFailuresProxy(0, remoteHost, remotePort); |
| networkFailuresProxies.add(proxy); |
| return proxy; |
| } |
| } |