blob: ffcbcb37b8a5467d0741c5c1a172cc89832c5a49 [file] [log] [blame]
/*
* Licensed to the Apache Software Foundation (ASF) under one or more
* contributor license agreements. See the NOTICE file distributed with
* this work for additional information regarding copyright ownership.
* The ASF licenses this file to You under the Apache License, Version 2.0
* (the "License"); you may not use this file except in compliance with
* the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package org.apache.camel.component.kafka.springboot;
import java.util.Map;
import java.util.concurrent.ExecutorService;
import javax.annotation.Generated;
import org.apache.camel.component.kafka.KafkaClientFactory;
import org.apache.camel.component.kafka.KafkaComponent;
import org.apache.camel.component.kafka.KafkaConfiguration;
import org.apache.camel.component.kafka.PollExceptionStrategy;
import org.apache.camel.component.kafka.PollOnError;
import org.apache.camel.component.kafka.SeekPolicy;
import org.apache.camel.component.kafka.consumer.KafkaManualCommitFactory;
import org.apache.camel.component.kafka.serde.KafkaHeaderDeserializer;
import org.apache.camel.component.kafka.serde.KafkaHeaderSerializer;
import org.apache.camel.spi.HeaderFilterStrategy;
import org.apache.camel.spi.StateRepository;
import org.apache.camel.spring.boot.ComponentConfigurationPropertiesCommon;
import org.apache.camel.support.jsse.SSLContextParameters;
import org.springframework.boot.context.properties.ConfigurationProperties;
/**
* Sent and receive messages to/from an Apache Kafka broker.
*
* Generated by camel-package-maven-plugin - do not edit this file!
*/
@Generated("org.apache.camel.springboot.maven.SpringBootAutoConfigurationMojo")
@ConfigurationProperties(prefix = "camel.component.kafka")
public class KafkaComponentConfiguration
extends
ComponentConfigurationPropertiesCommon {
/**
* Whether to enable auto configuration of the kafka component. This is
* enabled by default.
*/
private Boolean enabled;
/**
* Sets additional properties for either kafka consumer or kafka producer in
* case they can't be set directly on the camel configurations (e.g: new
* Kafka properties that are not reflected yet in Camel configurations), the
* properties have to be prefixed with additionalProperties.. E.g:
* additionalProperties.transactional.id=12345&additionalProperties.schema.registry.url=http://localhost:8811/avro
*/
private Map<String, Object> additionalProperties;
/**
* URL of the Kafka brokers to use. The format is host1:port1,host2:port2,
* and the list can be a subset of brokers or a VIP pointing to a subset of
* brokers. This option is known as bootstrap.servers in the Kafka
* documentation.
*/
private String brokers;
/**
* The client id is a user-specified string sent in each request to help
* trace calls. It should logically identify the application making the
* request.
*/
private String clientId;
/**
* Allows to pre-configure the Kafka component with common options that the
* endpoints will reuse. The option is a
* org.apache.camel.component.kafka.KafkaConfiguration type.
*/
private KafkaConfiguration configuration;
/**
* To use a custom HeaderFilterStrategy to filter header to and from Camel
* message. The option is a org.apache.camel.spi.HeaderFilterStrategy type.
*/
private HeaderFilterStrategy headerFilterStrategy;
/**
* The maximum amount of time in milliseconds to wait when reconnecting to a
* broker that has repeatedly failed to connect. If provided, the backoff
* per host will increase exponentially for each consecutive connection
* failure, up to this maximum. After calculating the backoff increase, 20%
* random jitter is added to avoid connection storms.
*/
private Integer reconnectBackoffMaxMs = 1000;
/**
* Timeout in milliseconds to wait gracefully for the consumer or producer
* to shutdown and terminate its worker threads.
*/
private Integer shutdownTimeout = 30000;
/**
* Whether to allow doing manual commits via KafkaManualCommit. If this
* option is enabled then an instance of KafkaManualCommit is stored on the
* Exchange message header, which allows end users to access this API and
* perform manual offset commits via the Kafka consumer.
*/
private Boolean allowManualCommit = false;
/**
* If true, periodically commit to ZooKeeper the offset of messages already
* fetched by the consumer. This committed offset will be used when the
* process fails as the position from which the new consumer will begin.
*/
private Boolean autoCommitEnable = true;
/**
* The frequency in ms that the consumer offsets are committed to zookeeper.
*/
private Integer autoCommitIntervalMs = 5000;
/**
* What to do when there is no initial offset in ZooKeeper or if an offset
* is out of range: earliest : automatically reset the offset to the
* earliest offset latest : automatically reset the offset to the latest
* offset fail: throw exception to the consumer
*/
private String autoOffsetReset = "latest";
/**
* This options controls what happens when a consumer is processing an
* exchange and it fails. If the option is false then the consumer continues
* to the next message and processes it. If the option is true then the
* consumer breaks out, and will seek back to offset of the message that
* caused a failure, and then re-attempt to process this message. However
* this can lead to endless processing of the same message if its bound to
* fail every time, eg a poison message. Therefore its recommended to deal
* with that for example by using Camel's error handler.
*/
private Boolean breakOnFirstError = false;
/**
* Allows for bridging the consumer to the Camel routing Error Handler,
* which mean any exceptions occurred while the consumer is trying to pickup
* incoming messages, or the likes, will now be processed as a message and
* handled by the routing Error Handler. By default the consumer will use
* the org.apache.camel.spi.ExceptionHandler to deal with exceptions, that
* will be logged at WARN or ERROR level and ignored.
*/
private Boolean bridgeErrorHandler = false;
/**
* Automatically check the CRC32 of the records consumed. This ensures no
* on-the-wire or on-disk corruption to the messages occurred. This check
* adds some overhead, so it may be disabled in cases seeking extreme
* performance.
*/
private Boolean checkCrcs = true;
/**
* The maximum time, in milliseconds, that the code will wait for a
* synchronous commit to complete. The option is a java.lang.Long type.
*/
private Long commitTimeoutMs = 5000L;
/**
* The configuration controls the maximum amount of time the client will
* wait for the response of a request. If the response is not received
* before the timeout elapses the client will resend the request if
* necessary or fail the request if retries are exhausted.
*/
private Integer consumerRequestTimeoutMs = 40000;
/**
* The number of consumers that connect to kafka server. Each consumer is
* run on a separate thread, that retrieves and process the incoming data.
*/
private Integer consumersCount = 1;
/**
* The maximum amount of data the server should return for a fetch request
* This is not an absolute maximum, if the first message in the first
* non-empty partition of the fetch is larger than this value, the message
* will still be returned to ensure that the consumer can make progress. The
* maximum message size accepted by the broker is defined via
* message.max.bytes (broker config) or max.message.bytes (topic config).
* Note that the consumer performs multiple fetches in parallel.
*/
private Integer fetchMaxBytes = 52428800;
/**
* The minimum amount of data the server should return for a fetch request.
* If insufficient data is available the request will wait for that much
* data to accumulate before answering the request.
*/
private Integer fetchMinBytes = 1;
/**
* The maximum amount of time the server will block before answering the
* fetch request if there isn't sufficient data to immediately satisfy
* fetch.min.bytes
*/
private Integer fetchWaitMaxMs = 500;
/**
* A string that uniquely identifies the group of consumer processes to
* which this consumer belongs. By setting the same group id multiple
* processes indicate that they are all part of the same consumer group.
* This option is required for consumers.
*/
private String groupId;
/**
* A unique identifier of the consumer instance provided by the end user.
* Only non-empty strings are permitted. If set, the consumer is treated as
* a static member, which means that only one instance with this ID is
* allowed in the consumer group at any time. This can be used in
* combination with a larger session timeout to avoid group rebalances
* caused by transient unavailability (e.g. process restarts). If not set,
* the consumer will join the group as a dynamic member, which is the
* traditional behavior.
*/
private String groupInstanceId;
/**
* To use a custom KafkaHeaderDeserializer to deserialize kafka headers
* values. The option is a
* org.apache.camel.component.kafka.serde.KafkaHeaderDeserializer type.
*/
private KafkaHeaderDeserializer headerDeserializer;
/**
* The expected time between heartbeats to the consumer coordinator when
* using Kafka's group management facilities. Heartbeats are used to ensure
* that the consumer's session stays active and to facilitate rebalancing
* when new consumers join or leave the group. The value must be set lower
* than session.timeout.ms, but typically should be set no higher than 1/3
* of that value. It can be adjusted even lower to control the expected time
* for normal rebalances.
*/
private Integer heartbeatIntervalMs = 3000;
/**
* Deserializer class for key that implements the Deserializer interface.
*/
private String keyDeserializer = "org.apache.kafka.common.serialization.StringDeserializer";
/**
* The maximum amount of data per-partition the server will return. The
* maximum total memory used for a request will be #partitions
* max.partition.fetch.bytes. This size must be at least as large as the
* maximum message size the server allows or else it is possible for the
* producer to send messages larger than the consumer can fetch. If that
* happens, the consumer can get stuck trying to fetch a large message on a
* certain partition.
*/
private Integer maxPartitionFetchBytes = 1048576;
/**
* The maximum delay between invocations of poll() when using consumer group
* management. This places an upper bound on the amount of time that the
* consumer can be idle before fetching more records. If poll() is not
* called before expiration of this timeout, then the consumer is considered
* failed and the group will rebalance in order to reassign the partitions
* to another member. The option is a java.lang.Long type.
*/
private Long maxPollIntervalMs;
/**
* The maximum number of records returned in a single call to poll()
*/
private Integer maxPollRecords = 500;
/**
* The offset repository to use in order to locally store the offset of each
* partition of the topic. Defining one will disable the autocommit. The
* option is a org.apache.camel.spi.StateRepository<java.lang.String,
* java.lang.String> type.
*/
private StateRepository<String, String> offsetRepository;
/**
* The class name of the partition assignment strategy that the client will
* use to distribute partition ownership amongst consumer instances when
* group management is used
*/
private String partitionAssignor = "org.apache.kafka.clients.consumer.RangeAssignor";
/**
* What to do if kafka threw an exception while polling for new messages.
* Will by default use the value from the component configuration unless an
* explicit value has been configured on the endpoint level. DISCARD will
* discard the message and continue to poll next message. ERROR_HANDLER will
* use Camel's error handler to process the exception, and afterwards
* continue to poll next message. RECONNECT will re-connect the consumer and
* try poll the message again RETRY will let the consumer retry polling the
* same message again STOP will stop the consumer (have to be manually
* started/restarted if the consumer should be able to consume messages
* again)
*/
private PollOnError pollOnError = PollOnError.ERROR_HANDLER;
/**
* The timeout used when polling the KafkaConsumer. The option is a
* java.lang.Long type.
*/
private Long pollTimeoutMs = 5000L;
/**
* Set if KafkaConsumer will read from beginning or end on startup:
* SeekPolicy.BEGINNING: read from beginning. SeekPolicy.END: read from end.
*/
private SeekPolicy seekTo;
/**
* The timeout used to detect failures when using Kafka's group management
* facilities.
*/
private Integer sessionTimeoutMs = 10000;
/**
* This enables the use of a specific Avro reader for use with the Confluent
* Platform schema registry and the
* io.confluent.kafka.serializers.KafkaAvroDeserializer. This option is only
* available in the Confluent Platform (not standard Apache Kafka)
*/
private Boolean specificAvroReader = false;
/**
* Whether the topic is a pattern (regular expression). This can be used to
* subscribe to dynamic number of topics matching the pattern.
*/
private Boolean topicIsPattern = false;
/**
* Deserializer class for value that implements the Deserializer interface.
*/
private String valueDeserializer = "org.apache.kafka.common.serialization.StringDeserializer";
/**
* The delay in millis seconds to wait before trying again to create the
* kafka consumer (kafka-client).
*/
private Long createConsumerBackoffInterval = 5000L;
/**
* Maximum attempts to create the kafka consumer (kafka-client), before
* eventually giving up and failing. Error during creating the consumer may
* be fatal due to invalid configuration and as such recovery is not
* possible. However, one part of the validation is DNS resolution of the
* bootstrap broker hostnames. This may be a temporary networking problem,
* and could potentially be recoverable. While other errors are fatal such
* as some invalid kafka configurations. Unfortunately kafka-client does not
* separate this kind of errors. Camel will by default retry forever, and
* therefore never give up. If you want to give up after many attempts then
* set this option and Camel will then when giving up terminate the
* consumer. You can manually restart the consumer by stopping and starting
* the route, to try again.
*/
private Integer createConsumerBackoffMaxAttempts;
/**
* Controls how to read messages written transactionally. If set to
* read_committed, consumer.poll() will only return transactional messages
* which have been committed. If set to read_uncommitted (the default),
* consumer.poll() will return all messages, even transactional messages
* which have been aborted. Non-transactional messages will be returned
* unconditionally in either mode. Messages will always be returned in
* offset order. Hence, in read_committed mode, consumer.poll() will only
* return messages up to the last stable offset (LSO), which is the one less
* than the offset of the first open transaction. In particular any messages
* appearing after messages belonging to ongoing transactions will be
* withheld until the relevant transaction has been completed. As a result,
* read_committed consumers will not be able to read up to the high
* watermark when there are in flight transactions. Further, when in
* read_committed the seekToEnd method will return the LSO
*/
private String isolationLevel = "read_uncommitted";
/**
* Factory to use for creating KafkaManualCommit instances. This allows to
* plugin a custom factory to create custom KafkaManualCommit instances in
* case special logic is needed when doing manual commits that deviates from
* the default implementation that comes out of the box. The option is a
* org.apache.camel.component.kafka.consumer.KafkaManualCommitFactory type.
*/
private KafkaManualCommitFactory kafkaManualCommitFactory;
/**
* To use a custom strategy with the consumer to control how to handle
* exceptions thrown from the Kafka broker while pooling messages. The
* option is a org.apache.camel.component.kafka.PollExceptionStrategy type.
*/
private PollExceptionStrategy pollExceptionStrategy;
/**
* The delay in millis seconds to wait before trying again to subscribe to
* the kafka broker.
*/
private Long subscribeConsumerBackoffInterval = 5000L;
/**
* Maximum number the kafka consumer will attempt to subscribe to the kafka
* broker, before eventually giving up and failing. Error during subscribing
* the consumer to the kafka topic could be temporary errors due to network
* issues, and could potentially be recoverable. Camel will by default retry
* forever, and therefore never give up. If you want to give up after many
* attempts then set this option and Camel will then when giving up
* terminate the consumer. You can manually restart the consumer by stopping
* and starting the route, to try again.
*/
private Integer subscribeConsumerBackoffMaxAttempts;
/**
* If this feature is enabled and a single element of a batch is an Exchange
* or Message, the producer will generate individual kafka header values for
* it by using the batch Message to determine the values. Normal behaviour
* consists in always using the same header values (which are determined by
* the parent Exchange which contains the Iterable or Iterator).
*/
private Boolean batchWithIndividualHeaders = false;
/**
* The total bytes of memory the producer can use to buffer records waiting
* to be sent to the server. If records are sent faster than they can be
* delivered to the server the producer will either block or throw an
* exception based on the preference specified by block.on.buffer.full.This
* setting should correspond roughly to the total memory the producer will
* use, but is not a hard bound since not all memory the producer uses is
* used for buffering. Some additional memory will be used for compression
* (if compression is enabled) as well as for maintaining in-flight
* requests.
*/
private Integer bufferMemorySize = 33554432;
/**
* This parameter allows you to specify the compression codec for all data
* generated by this producer. Valid values are none, gzip and snappy.
*/
private String compressionCodec = "none";
/**
* Close idle connections after the number of milliseconds specified by this
* config.
*/
private Integer connectionMaxIdleMs = 540000;
/**
* An upper bound on the time to report success or failure after a call to
* send() returns. This limits the total time that a record will be delayed
* prior to sending, the time to await acknowledgement from the broker (if
* expected), and the time allowed for retriable send failures.
*/
private Integer deliveryTimeoutMs = 120000;
/**
* If set to 'true' the producer will ensure that exactly one copy of each
* message is written in the stream. If 'false', producer retries may write
* duplicates of the retried message in the stream. If set to true this
* option will require max.in.flight.requests.per.connection to be set to 1
* and retries cannot be zero and additionally acks must be set to 'all'.
*/
private Boolean enableIdempotence = false;
/**
* To use a custom KafkaHeaderSerializer to serialize kafka headers values.
* The option is a
* org.apache.camel.component.kafka.serde.KafkaHeaderSerializer type.
*/
private KafkaHeaderSerializer headerSerializer;
/**
* The record key (or null if no key is specified). If this option has been
* configured then it take precedence over header KafkaConstants#KEY
*/
private String key;
/**
* The serializer class for keys (defaults to the same as for messages if
* nothing is given).
*/
private String keySerializer = "org.apache.kafka.common.serialization.StringSerializer";
/**
* Whether the producer should be started lazy (on the first message). By
* starting lazy you can use this to allow CamelContext and routes to
* startup in situations where a producer may otherwise fail during starting
* and cause the route to fail being started. By deferring this startup to
* be lazy then the startup failure can be handled during routing messages
* via Camel's routing error handlers. Beware that when the first message is
* processed then creating and starting the producer may take a little time
* and prolong the total processing time of the processing.
*/
private Boolean lazyStartProducer = false;
/**
* The producer groups together any records that arrive in between request
* transmissions into a single batched request. Normally this occurs only
* under load when records arrive faster than they can be sent out. However
* in some circumstances the client may want to reduce the number of
* requests even under moderate load. This setting accomplishes this by
* adding a small amount of artificial delay that is, rather than
* immediately sending out a record the producer will wait for up to the
* given delay to allow other records to be sent so that the sends can be
* batched together. This can be thought of as analogous to Nagle's
* algorithm in TCP. This setting gives the upper bound on the delay for
* batching: once we get batch.size worth of records for a partition it will
* be sent immediately regardless of this setting, however if we have fewer
* than this many bytes accumulated for this partition we will 'linger' for
* the specified time waiting for more records to show up. This setting
* defaults to 0 (i.e. no delay). Setting linger.ms=5, for example, would
* have the effect of reducing the number of requests sent but would add up
* to 5ms of latency to records sent in the absense of load.
*/
private Integer lingerMs = 0;
/**
* The configuration controls how long the KafkaProducer's send(),
* partitionsFor(), initTransactions(), sendOffsetsToTransaction(),
* commitTransaction() and abortTransaction() methods will block. For send()
* this timeout bounds the total time waiting for both metadata fetch and
* buffer allocation (blocking in the user-supplied serializers or
* partitioner is not counted against this timeout). For partitionsFor()
* this timeout bounds the time spent waiting for metadata if it is
* unavailable. The transaction-related methods always block, but may
* timeout if the transaction coordinator could not be discovered or did not
* respond within the timeout.
*/
private Integer maxBlockMs = 60000;
/**
* The maximum number of unacknowledged requests the client will send on a
* single connection before blocking. Note that if this setting is set to be
* greater than 1 and there are failed sends, there is a risk of message
* re-ordering due to retries (i.e., if retries are enabled).
*/
private Integer maxInFlightRequest = 5;
/**
* The maximum size of a request. This is also effectively a cap on the
* maximum record size. Note that the server has its own cap on record size
* which may be different from this. This setting will limit the number of
* record batches the producer will send in a single request to avoid
* sending huge requests.
*/
private Integer maxRequestSize = 1048576;
/**
* The period of time in milliseconds after which we force a refresh of
* metadata even if we haven't seen any partition leadership changes to
* proactively discover any new brokers or partitions.
*/
private Integer metadataMaxAgeMs = 300000;
/**
* A list of classes to use as metrics reporters. Implementing the
* MetricReporter interface allows plugging in classes that will be notified
* of new metric creation. The JmxReporter is always included to register
* JMX statistics.
*/
private String metricReporters;
/**
* The window of time a metrics sample is computed over.
*/
private Integer metricsSampleWindowMs = 30000;
/**
* The number of samples maintained to compute metrics.
*/
private Integer noOfMetricsSample = 2;
/**
* The partitioner class for partitioning messages amongst sub-topics. The
* default partitioner is based on the hash of the key.
*/
private String partitioner = "org.apache.kafka.clients.producer.internals.DefaultPartitioner";
/**
* The partition to which the record will be sent (or null if no partition
* was specified). If this option has been configured then it take
* precedence over header KafkaConstants#PARTITION_KEY
*/
private Integer partitionKey;
/**
* The producer will attempt to batch records together into fewer requests
* whenever multiple records are being sent to the same partition. This
* helps performance on both the client and the server. This configuration
* controls the default batch size in bytes. No attempt will be made to
* batch records larger than this size.Requests sent to brokers will contain
* multiple batches, one for each partition with data available to be sent.A
* small batch size will make batching less common and may reduce throughput
* (a batch size of zero will disable batching entirely). A very large batch
* size may use memory a bit more wastefully as we will always allocate a
* buffer of the specified batch size in anticipation of additional records.
*/
private Integer producerBatchSize = 16384;
/**
* The maximum number of unsent messages that can be queued up the producer
* when using async mode before either the producer must be blocked or data
* must be dropped.
*/
private Integer queueBufferingMaxMessages = 10000;
/**
* The size of the TCP receive buffer (SO_RCVBUF) to use when reading data.
*/
private Integer receiveBufferBytes = 65536;
/**
* The amount of time to wait before attempting to reconnect to a given
* host. This avoids repeatedly connecting to a host in a tight loop. This
* backoff applies to all requests sent by the consumer to the broker.
*/
private Integer reconnectBackoffMs = 50;
/**
* Whether the producer should store the RecordMetadata results from sending
* to Kafka. The results are stored in a List containing the RecordMetadata
* metadata's. The list is stored on a header with the key
* KafkaConstants#KAFKA_RECORDMETA
*/
private Boolean recordMetadata = true;
/**
* The number of acknowledgments the producer requires the leader to have
* received before considering a request complete. This controls the
* durability of records that are sent. The following settings are common:
* acks=0 If set to zero then the producer will not wait for any
* acknowledgment from the server at all. The record will be immediately
* added to the socket buffer and considered sent. No guarantee can be made
* that the server has received the record in this case, and the retries
* configuration will not take effect (as the client won't generally know of
* any failures). The offset given back for each record will always be set
* to -1. acks=1 This will mean the leader will write the record to its
* local log but will respond without awaiting full acknowledgement from all
* followers. In this case should the leader fail immediately after
* acknowledging the record but before the followers have replicated it then
* the record will be lost. acks=all This means the leader will wait for the
* full set of in-sync replicas to acknowledge the record. This guarantees
* that the record will not be lost as long as at least one in-sync replica
* remains alive. This is the strongest available guarantee.
*/
private String requestRequiredAcks = "1";
/**
* The amount of time the broker will wait trying to meet the
* request.required.acks requirement before sending back an error to the
* client.
*/
private Integer requestTimeoutMs = 30000;
/**
* Setting a value greater than zero will cause the client to resend any
* record whose send fails with a potentially transient error. Note that
* this retry is no different than if the client resent the record upon
* receiving the error. Allowing retries will potentially change the
* ordering of records because if two records are sent to a single
* partition, and the first fails and is retried but the second succeeds,
* then the second record may appear first.
*/
private Integer retries = 0;
/**
* Before each retry, the producer refreshes the metadata of relevant topics
* to see if a new leader has been elected. Since leader election takes a
* bit of time, this property specifies the amount of time that the producer
* waits before refreshing the metadata.
*/
private Integer retryBackoffMs = 100;
/**
* Socket write buffer size
*/
private Integer sendBufferBytes = 131072;
/**
* The serializer class for messages.
*/
private String valueSerializer = "org.apache.kafka.common.serialization.StringSerializer";
/**
* To use a custom worker pool for continue routing Exchange after kafka
* server has acknowledge the message that was sent to it from KafkaProducer
* using asynchronous non-blocking processing. If using this option then you
* must handle the lifecycle of the thread pool to shut the pool down when
* no longer needed. The option is a java.util.concurrent.ExecutorService
* type.
*/
private ExecutorService workerPool;
/**
* Number of core threads for the worker pool for continue routing Exchange
* after kafka server has acknowledge the message that was sent to it from
* KafkaProducer using asynchronous non-blocking processing.
*/
private Integer workerPoolCoreSize = 10;
/**
* Maximum number of threads for the worker pool for continue routing
* Exchange after kafka server has acknowledge the message that was sent to
* it from KafkaProducer using asynchronous non-blocking processing.
*/
private Integer workerPoolMaxSize = 20;
/**
* Whether autowiring is enabled. This is used for automatic autowiring
* options (the option must be marked as autowired) by looking up in the
* registry to find if there is a single instance of matching type, which
* then gets configured on the component. This can be used for automatic
* configuring JDBC data sources, JMS connection factories, AWS Clients,
* etc.
*/
private Boolean autowiredEnabled = true;
/**
* Factory to use for creating
* org.apache.kafka.clients.consumer.KafkaConsumer and
* org.apache.kafka.clients.producer.KafkaProducer instances. This allows to
* configure a custom factory to create instances with logic that extends
* the vanilla Kafka clients. The option is a
* org.apache.camel.component.kafka.KafkaClientFactory type.
*/
private KafkaClientFactory kafkaClientFactory;
/**
* Sets whether synchronous processing should be strictly used
*/
private Boolean synchronous = false;
/**
* URL of the Confluent Platform schema registry servers to use. The format
* is host1:port1,host2:port2. This is known as schema.registry.url in the
* Confluent Platform documentation. This option is only available in the
* Confluent Platform (not standard Apache Kafka)
*/
private String schemaRegistryURL;
/**
* Sets interceptors for producer or consumers. Producer interceptors have
* to be classes implementing
* org.apache.kafka.clients.producer.ProducerInterceptor Consumer
* interceptors have to be classes implementing
* org.apache.kafka.clients.consumer.ConsumerInterceptor Note that if you
* use Producer interceptor on a consumer it will throw a class cast
* exception in runtime
*/
private String interceptorClasses;
/**
* Login thread sleep time between refresh attempts.
*/
private Integer kerberosBeforeReloginMinTime = 60000;
/**
* Kerberos kinit command path. Default is /usr/bin/kinit
*/
private String kerberosInitCmd = "/usr/bin/kinit";
/**
* A list of rules for mapping from principal names to short names
* (typically operating system usernames). The rules are evaluated in order
* and the first rule that matches a principal name is used to map it to a
* short name. Any later rules in the list are ignored. By default,
* principal names of the form {username}/{hostname}{REALM} are mapped to
* {username}. For more details on the format please see the security
* authorization and acls documentation (at the Apache Kafka project).
* Multiple values can be separated by comma
*/
private String kerberosPrincipalToLocalRules = "DEFAULT";
/**
* Percentage of random jitter added to the renewal time.
*/
private Double kerberosRenewJitter;
/**
* Login thread will sleep until the specified window factor of time from
* last refresh to ticket's expiry has been reached, at which time it will
* try to renew the ticket.
*/
private Double kerberosRenewWindowFactor;
/**
* Expose the kafka sasl.jaas.config parameter Example:
* org.apache.kafka.common.security.plain.PlainLoginModule required
* username=USERNAME password=PASSWORD;
*/
private String saslJaasConfig;
/**
* The Kerberos principal name that Kafka runs as. This can be defined
* either in Kafka's JAAS config or in Kafka's config.
*/
private String saslKerberosServiceName;
/**
* The Simple Authentication and Security Layer (SASL) Mechanism used. For
* the valid values see
* http://www.iana.org/assignments/sasl-mechanisms/sasl-mechanisms.xhtml
*/
private String saslMechanism = "GSSAPI";
/**
* Protocol used to communicate with brokers. SASL_PLAINTEXT, PLAINTEXT and
* SSL are supported
*/
private String securityProtocol = "PLAINTEXT";
/**
* A list of cipher suites. This is a named combination of authentication,
* encryption, MAC and key exchange algorithm used to negotiate the security
* settings for a network connection using TLS or SSL network protocol. By
* default all the available cipher suites are supported.
*/
private String sslCipherSuites;
/**
* SSL configuration using a Camel SSLContextParameters object. If
* configured it's applied before the other SSL endpoint parameters. NOTE:
* Kafka only supports loading keystore from file locations, so prefix the
* location with file: in the KeyStoreParameters.resource option. The option
* is a org.apache.camel.support.jsse.SSLContextParameters type.
*/
private SSLContextParameters sslContextParameters;
/**
* The list of protocols enabled for SSL connections. The default is
* TLSv1.2,TLSv1.3 when running with Java 11 or newer, TLSv1.2 otherwise.
* With the default value for Java 11, clients and servers will prefer
* TLSv1.3 if both support it and fallback to TLSv1.2 otherwise (assuming
* both support at least TLSv1.2). This default should be fine for most
* cases. Also see the config documentation for SslProtocol.
*/
private String sslEnabledProtocols;
/**
* The endpoint identification algorithm to validate server hostname using
* server certificate. Use none or false to disable server hostname
* verification.
*/
private String sslEndpointAlgorithm = "https";
/**
* The algorithm used by key manager factory for SSL connections. Default
* value is the key manager factory algorithm configured for the Java
* Virtual Machine.
*/
private String sslKeymanagerAlgorithm = "SunX509";
/**
* The password of the private key in the key store file or the PEM key
* specified in sslKeystoreKey. This is required for clients only if two-way
* authentication is configured.
*/
private String sslKeyPassword;
/**
* The location of the key store file. This is optional for client and can
* be used for two-way authentication for client.
*/
private String sslKeystoreLocation;
/**
* The store password for the key store file. This is optional for client
* and only needed if sslKeystoreLocation' is configured. Key store password
* is not supported for PEM format.
*/
private String sslKeystorePassword;
/**
* The file format of the key store file. This is optional for client.
* Default value is JKS
*/
private String sslKeystoreType = "JKS";
/**
* The SSL protocol used to generate the SSLContext. The default is TLSv1.3
* when running with Java 11 or newer, TLSv1.2 otherwise. This value should
* be fine for most use cases. Allowed values in recent JVMs are TLSv1.2 and
* TLSv1.3. TLS, TLSv1.1, SSL, SSLv2 and SSLv3 may be supported in older
* JVMs, but their usage is discouraged due to known security
* vulnerabilities. With the default value for this config and
* sslEnabledProtocols, clients will downgrade to TLSv1.2 if the server does
* not support TLSv1.3. If this config is set to TLSv1.2, clients will not
* use TLSv1.3 even if it is one of the values in sslEnabledProtocols and
* the server only supports TLSv1.3.
*/
private String sslProtocol;
/**
* The name of the security provider used for SSL connections. Default value
* is the default security provider of the JVM.
*/
private String sslProvider;
/**
* The algorithm used by trust manager factory for SSL connections. Default
* value is the trust manager factory algorithm configured for the Java
* Virtual Machine.
*/
private String sslTrustmanagerAlgorithm = "PKIX";
/**
* The location of the trust store file.
*/
private String sslTruststoreLocation;
/**
* The password for the trust store file. If a password is not set, trust
* store file configured will still be used, but integrity checking is
* disabled. Trust store password is not supported for PEM format.
*/
private String sslTruststorePassword;
/**
* The file format of the trust store file. Default value is JKS.
*/
private String sslTruststoreType = "JKS";
/**
* Enable usage of global SSL context parameters.
*/
private Boolean useGlobalSslContextParameters = false;
public Map<String, Object> getAdditionalProperties() {
return additionalProperties;
}
public void setAdditionalProperties(Map<String, Object> additionalProperties) {
this.additionalProperties = additionalProperties;
}
public String getBrokers() {
return brokers;
}
public void setBrokers(String brokers) {
this.brokers = brokers;
}
public String getClientId() {
return clientId;
}
public void setClientId(String clientId) {
this.clientId = clientId;
}
public KafkaConfiguration getConfiguration() {
return configuration;
}
public void setConfiguration(KafkaConfiguration configuration) {
this.configuration = configuration;
}
public HeaderFilterStrategy getHeaderFilterStrategy() {
return headerFilterStrategy;
}
public void setHeaderFilterStrategy(
HeaderFilterStrategy headerFilterStrategy) {
this.headerFilterStrategy = headerFilterStrategy;
}
public Integer getReconnectBackoffMaxMs() {
return reconnectBackoffMaxMs;
}
public void setReconnectBackoffMaxMs(Integer reconnectBackoffMaxMs) {
this.reconnectBackoffMaxMs = reconnectBackoffMaxMs;
}
public Integer getShutdownTimeout() {
return shutdownTimeout;
}
public void setShutdownTimeout(Integer shutdownTimeout) {
this.shutdownTimeout = shutdownTimeout;
}
public Boolean getAllowManualCommit() {
return allowManualCommit;
}
public void setAllowManualCommit(Boolean allowManualCommit) {
this.allowManualCommit = allowManualCommit;
}
public Boolean getAutoCommitEnable() {
return autoCommitEnable;
}
public void setAutoCommitEnable(Boolean autoCommitEnable) {
this.autoCommitEnable = autoCommitEnable;
}
public Integer getAutoCommitIntervalMs() {
return autoCommitIntervalMs;
}
public void setAutoCommitIntervalMs(Integer autoCommitIntervalMs) {
this.autoCommitIntervalMs = autoCommitIntervalMs;
}
public String getAutoOffsetReset() {
return autoOffsetReset;
}
public void setAutoOffsetReset(String autoOffsetReset) {
this.autoOffsetReset = autoOffsetReset;
}
public Boolean getBreakOnFirstError() {
return breakOnFirstError;
}
public void setBreakOnFirstError(Boolean breakOnFirstError) {
this.breakOnFirstError = breakOnFirstError;
}
public Boolean getBridgeErrorHandler() {
return bridgeErrorHandler;
}
public void setBridgeErrorHandler(Boolean bridgeErrorHandler) {
this.bridgeErrorHandler = bridgeErrorHandler;
}
public Boolean getCheckCrcs() {
return checkCrcs;
}
public void setCheckCrcs(Boolean checkCrcs) {
this.checkCrcs = checkCrcs;
}
public Long getCommitTimeoutMs() {
return commitTimeoutMs;
}
public void setCommitTimeoutMs(Long commitTimeoutMs) {
this.commitTimeoutMs = commitTimeoutMs;
}
public Integer getConsumerRequestTimeoutMs() {
return consumerRequestTimeoutMs;
}
public void setConsumerRequestTimeoutMs(Integer consumerRequestTimeoutMs) {
this.consumerRequestTimeoutMs = consumerRequestTimeoutMs;
}
public Integer getConsumersCount() {
return consumersCount;
}
public void setConsumersCount(Integer consumersCount) {
this.consumersCount = consumersCount;
}
public Integer getFetchMaxBytes() {
return fetchMaxBytes;
}
public void setFetchMaxBytes(Integer fetchMaxBytes) {
this.fetchMaxBytes = fetchMaxBytes;
}
public Integer getFetchMinBytes() {
return fetchMinBytes;
}
public void setFetchMinBytes(Integer fetchMinBytes) {
this.fetchMinBytes = fetchMinBytes;
}
public Integer getFetchWaitMaxMs() {
return fetchWaitMaxMs;
}
public void setFetchWaitMaxMs(Integer fetchWaitMaxMs) {
this.fetchWaitMaxMs = fetchWaitMaxMs;
}
public String getGroupId() {
return groupId;
}
public void setGroupId(String groupId) {
this.groupId = groupId;
}
public String getGroupInstanceId() {
return groupInstanceId;
}
public void setGroupInstanceId(String groupInstanceId) {
this.groupInstanceId = groupInstanceId;
}
public KafkaHeaderDeserializer getHeaderDeserializer() {
return headerDeserializer;
}
public void setHeaderDeserializer(KafkaHeaderDeserializer headerDeserializer) {
this.headerDeserializer = headerDeserializer;
}
public Integer getHeartbeatIntervalMs() {
return heartbeatIntervalMs;
}
public void setHeartbeatIntervalMs(Integer heartbeatIntervalMs) {
this.heartbeatIntervalMs = heartbeatIntervalMs;
}
public String getKeyDeserializer() {
return keyDeserializer;
}
public void setKeyDeserializer(String keyDeserializer) {
this.keyDeserializer = keyDeserializer;
}
public Integer getMaxPartitionFetchBytes() {
return maxPartitionFetchBytes;
}
public void setMaxPartitionFetchBytes(Integer maxPartitionFetchBytes) {
this.maxPartitionFetchBytes = maxPartitionFetchBytes;
}
public Long getMaxPollIntervalMs() {
return maxPollIntervalMs;
}
public void setMaxPollIntervalMs(Long maxPollIntervalMs) {
this.maxPollIntervalMs = maxPollIntervalMs;
}
public Integer getMaxPollRecords() {
return maxPollRecords;
}
public void setMaxPollRecords(Integer maxPollRecords) {
this.maxPollRecords = maxPollRecords;
}
public StateRepository<String, String> getOffsetRepository() {
return offsetRepository;
}
public void setOffsetRepository(
StateRepository<String, String> offsetRepository) {
this.offsetRepository = offsetRepository;
}
public String getPartitionAssignor() {
return partitionAssignor;
}
public void setPartitionAssignor(String partitionAssignor) {
this.partitionAssignor = partitionAssignor;
}
public PollOnError getPollOnError() {
return pollOnError;
}
public void setPollOnError(PollOnError pollOnError) {
this.pollOnError = pollOnError;
}
public Long getPollTimeoutMs() {
return pollTimeoutMs;
}
public void setPollTimeoutMs(Long pollTimeoutMs) {
this.pollTimeoutMs = pollTimeoutMs;
}
public SeekPolicy getSeekTo() {
return seekTo;
}
public void setSeekTo(SeekPolicy seekTo) {
this.seekTo = seekTo;
}
public Integer getSessionTimeoutMs() {
return sessionTimeoutMs;
}
public void setSessionTimeoutMs(Integer sessionTimeoutMs) {
this.sessionTimeoutMs = sessionTimeoutMs;
}
public Boolean getSpecificAvroReader() {
return specificAvroReader;
}
public void setSpecificAvroReader(Boolean specificAvroReader) {
this.specificAvroReader = specificAvroReader;
}
public Boolean getTopicIsPattern() {
return topicIsPattern;
}
public void setTopicIsPattern(Boolean topicIsPattern) {
this.topicIsPattern = topicIsPattern;
}
public String getValueDeserializer() {
return valueDeserializer;
}
public void setValueDeserializer(String valueDeserializer) {
this.valueDeserializer = valueDeserializer;
}
public Long getCreateConsumerBackoffInterval() {
return createConsumerBackoffInterval;
}
public void setCreateConsumerBackoffInterval(
Long createConsumerBackoffInterval) {
this.createConsumerBackoffInterval = createConsumerBackoffInterval;
}
public Integer getCreateConsumerBackoffMaxAttempts() {
return createConsumerBackoffMaxAttempts;
}
public void setCreateConsumerBackoffMaxAttempts(
Integer createConsumerBackoffMaxAttempts) {
this.createConsumerBackoffMaxAttempts = createConsumerBackoffMaxAttempts;
}
public String getIsolationLevel() {
return isolationLevel;
}
public void setIsolationLevel(String isolationLevel) {
this.isolationLevel = isolationLevel;
}
public KafkaManualCommitFactory getKafkaManualCommitFactory() {
return kafkaManualCommitFactory;
}
public void setKafkaManualCommitFactory(
KafkaManualCommitFactory kafkaManualCommitFactory) {
this.kafkaManualCommitFactory = kafkaManualCommitFactory;
}
public PollExceptionStrategy getPollExceptionStrategy() {
return pollExceptionStrategy;
}
public void setPollExceptionStrategy(
PollExceptionStrategy pollExceptionStrategy) {
this.pollExceptionStrategy = pollExceptionStrategy;
}
public Long getSubscribeConsumerBackoffInterval() {
return subscribeConsumerBackoffInterval;
}
public void setSubscribeConsumerBackoffInterval(
Long subscribeConsumerBackoffInterval) {
this.subscribeConsumerBackoffInterval = subscribeConsumerBackoffInterval;
}
public Integer getSubscribeConsumerBackoffMaxAttempts() {
return subscribeConsumerBackoffMaxAttempts;
}
public void setSubscribeConsumerBackoffMaxAttempts(
Integer subscribeConsumerBackoffMaxAttempts) {
this.subscribeConsumerBackoffMaxAttempts = subscribeConsumerBackoffMaxAttempts;
}
public Boolean getBatchWithIndividualHeaders() {
return batchWithIndividualHeaders;
}
public void setBatchWithIndividualHeaders(Boolean batchWithIndividualHeaders) {
this.batchWithIndividualHeaders = batchWithIndividualHeaders;
}
public Integer getBufferMemorySize() {
return bufferMemorySize;
}
public void setBufferMemorySize(Integer bufferMemorySize) {
this.bufferMemorySize = bufferMemorySize;
}
public String getCompressionCodec() {
return compressionCodec;
}
public void setCompressionCodec(String compressionCodec) {
this.compressionCodec = compressionCodec;
}
public Integer getConnectionMaxIdleMs() {
return connectionMaxIdleMs;
}
public void setConnectionMaxIdleMs(Integer connectionMaxIdleMs) {
this.connectionMaxIdleMs = connectionMaxIdleMs;
}
public Integer getDeliveryTimeoutMs() {
return deliveryTimeoutMs;
}
public void setDeliveryTimeoutMs(Integer deliveryTimeoutMs) {
this.deliveryTimeoutMs = deliveryTimeoutMs;
}
public Boolean getEnableIdempotence() {
return enableIdempotence;
}
public void setEnableIdempotence(Boolean enableIdempotence) {
this.enableIdempotence = enableIdempotence;
}
public KafkaHeaderSerializer getHeaderSerializer() {
return headerSerializer;
}
public void setHeaderSerializer(KafkaHeaderSerializer headerSerializer) {
this.headerSerializer = headerSerializer;
}
public String getKey() {
return key;
}
public void setKey(String key) {
this.key = key;
}
public String getKeySerializer() {
return keySerializer;
}
public void setKeySerializer(String keySerializer) {
this.keySerializer = keySerializer;
}
public Boolean getLazyStartProducer() {
return lazyStartProducer;
}
public void setLazyStartProducer(Boolean lazyStartProducer) {
this.lazyStartProducer = lazyStartProducer;
}
public Integer getLingerMs() {
return lingerMs;
}
public void setLingerMs(Integer lingerMs) {
this.lingerMs = lingerMs;
}
public Integer getMaxBlockMs() {
return maxBlockMs;
}
public void setMaxBlockMs(Integer maxBlockMs) {
this.maxBlockMs = maxBlockMs;
}
public Integer getMaxInFlightRequest() {
return maxInFlightRequest;
}
public void setMaxInFlightRequest(Integer maxInFlightRequest) {
this.maxInFlightRequest = maxInFlightRequest;
}
public Integer getMaxRequestSize() {
return maxRequestSize;
}
public void setMaxRequestSize(Integer maxRequestSize) {
this.maxRequestSize = maxRequestSize;
}
public Integer getMetadataMaxAgeMs() {
return metadataMaxAgeMs;
}
public void setMetadataMaxAgeMs(Integer metadataMaxAgeMs) {
this.metadataMaxAgeMs = metadataMaxAgeMs;
}
public String getMetricReporters() {
return metricReporters;
}
public void setMetricReporters(String metricReporters) {
this.metricReporters = metricReporters;
}
public Integer getMetricsSampleWindowMs() {
return metricsSampleWindowMs;
}
public void setMetricsSampleWindowMs(Integer metricsSampleWindowMs) {
this.metricsSampleWindowMs = metricsSampleWindowMs;
}
public Integer getNoOfMetricsSample() {
return noOfMetricsSample;
}
public void setNoOfMetricsSample(Integer noOfMetricsSample) {
this.noOfMetricsSample = noOfMetricsSample;
}
public String getPartitioner() {
return partitioner;
}
public void setPartitioner(String partitioner) {
this.partitioner = partitioner;
}
public Integer getPartitionKey() {
return partitionKey;
}
public void setPartitionKey(Integer partitionKey) {
this.partitionKey = partitionKey;
}
public Integer getProducerBatchSize() {
return producerBatchSize;
}
public void setProducerBatchSize(Integer producerBatchSize) {
this.producerBatchSize = producerBatchSize;
}
public Integer getQueueBufferingMaxMessages() {
return queueBufferingMaxMessages;
}
public void setQueueBufferingMaxMessages(Integer queueBufferingMaxMessages) {
this.queueBufferingMaxMessages = queueBufferingMaxMessages;
}
public Integer getReceiveBufferBytes() {
return receiveBufferBytes;
}
public void setReceiveBufferBytes(Integer receiveBufferBytes) {
this.receiveBufferBytes = receiveBufferBytes;
}
public Integer getReconnectBackoffMs() {
return reconnectBackoffMs;
}
public void setReconnectBackoffMs(Integer reconnectBackoffMs) {
this.reconnectBackoffMs = reconnectBackoffMs;
}
public Boolean getRecordMetadata() {
return recordMetadata;
}
public void setRecordMetadata(Boolean recordMetadata) {
this.recordMetadata = recordMetadata;
}
public String getRequestRequiredAcks() {
return requestRequiredAcks;
}
public void setRequestRequiredAcks(String requestRequiredAcks) {
this.requestRequiredAcks = requestRequiredAcks;
}
public Integer getRequestTimeoutMs() {
return requestTimeoutMs;
}
public void setRequestTimeoutMs(Integer requestTimeoutMs) {
this.requestTimeoutMs = requestTimeoutMs;
}
public Integer getRetries() {
return retries;
}
public void setRetries(Integer retries) {
this.retries = retries;
}
public Integer getRetryBackoffMs() {
return retryBackoffMs;
}
public void setRetryBackoffMs(Integer retryBackoffMs) {
this.retryBackoffMs = retryBackoffMs;
}
public Integer getSendBufferBytes() {
return sendBufferBytes;
}
public void setSendBufferBytes(Integer sendBufferBytes) {
this.sendBufferBytes = sendBufferBytes;
}
public String getValueSerializer() {
return valueSerializer;
}
public void setValueSerializer(String valueSerializer) {
this.valueSerializer = valueSerializer;
}
public ExecutorService getWorkerPool() {
return workerPool;
}
public void setWorkerPool(ExecutorService workerPool) {
this.workerPool = workerPool;
}
public Integer getWorkerPoolCoreSize() {
return workerPoolCoreSize;
}
public void setWorkerPoolCoreSize(Integer workerPoolCoreSize) {
this.workerPoolCoreSize = workerPoolCoreSize;
}
public Integer getWorkerPoolMaxSize() {
return workerPoolMaxSize;
}
public void setWorkerPoolMaxSize(Integer workerPoolMaxSize) {
this.workerPoolMaxSize = workerPoolMaxSize;
}
public Boolean getAutowiredEnabled() {
return autowiredEnabled;
}
public void setAutowiredEnabled(Boolean autowiredEnabled) {
this.autowiredEnabled = autowiredEnabled;
}
public KafkaClientFactory getKafkaClientFactory() {
return kafkaClientFactory;
}
public void setKafkaClientFactory(KafkaClientFactory kafkaClientFactory) {
this.kafkaClientFactory = kafkaClientFactory;
}
public Boolean getSynchronous() {
return synchronous;
}
public void setSynchronous(Boolean synchronous) {
this.synchronous = synchronous;
}
public String getSchemaRegistryURL() {
return schemaRegistryURL;
}
public void setSchemaRegistryURL(String schemaRegistryURL) {
this.schemaRegistryURL = schemaRegistryURL;
}
public String getInterceptorClasses() {
return interceptorClasses;
}
public void setInterceptorClasses(String interceptorClasses) {
this.interceptorClasses = interceptorClasses;
}
public Integer getKerberosBeforeReloginMinTime() {
return kerberosBeforeReloginMinTime;
}
public void setKerberosBeforeReloginMinTime(
Integer kerberosBeforeReloginMinTime) {
this.kerberosBeforeReloginMinTime = kerberosBeforeReloginMinTime;
}
public String getKerberosInitCmd() {
return kerberosInitCmd;
}
public void setKerberosInitCmd(String kerberosInitCmd) {
this.kerberosInitCmd = kerberosInitCmd;
}
public String getKerberosPrincipalToLocalRules() {
return kerberosPrincipalToLocalRules;
}
public void setKerberosPrincipalToLocalRules(
String kerberosPrincipalToLocalRules) {
this.kerberosPrincipalToLocalRules = kerberosPrincipalToLocalRules;
}
public Double getKerberosRenewJitter() {
return kerberosRenewJitter;
}
public void setKerberosRenewJitter(Double kerberosRenewJitter) {
this.kerberosRenewJitter = kerberosRenewJitter;
}
public Double getKerberosRenewWindowFactor() {
return kerberosRenewWindowFactor;
}
public void setKerberosRenewWindowFactor(Double kerberosRenewWindowFactor) {
this.kerberosRenewWindowFactor = kerberosRenewWindowFactor;
}
public String getSaslJaasConfig() {
return saslJaasConfig;
}
public void setSaslJaasConfig(String saslJaasConfig) {
this.saslJaasConfig = saslJaasConfig;
}
public String getSaslKerberosServiceName() {
return saslKerberosServiceName;
}
public void setSaslKerberosServiceName(String saslKerberosServiceName) {
this.saslKerberosServiceName = saslKerberosServiceName;
}
public String getSaslMechanism() {
return saslMechanism;
}
public void setSaslMechanism(String saslMechanism) {
this.saslMechanism = saslMechanism;
}
public String getSecurityProtocol() {
return securityProtocol;
}
public void setSecurityProtocol(String securityProtocol) {
this.securityProtocol = securityProtocol;
}
public String getSslCipherSuites() {
return sslCipherSuites;
}
public void setSslCipherSuites(String sslCipherSuites) {
this.sslCipherSuites = sslCipherSuites;
}
public SSLContextParameters getSslContextParameters() {
return sslContextParameters;
}
public void setSslContextParameters(
SSLContextParameters sslContextParameters) {
this.sslContextParameters = sslContextParameters;
}
public String getSslEnabledProtocols() {
return sslEnabledProtocols;
}
public void setSslEnabledProtocols(String sslEnabledProtocols) {
this.sslEnabledProtocols = sslEnabledProtocols;
}
public String getSslEndpointAlgorithm() {
return sslEndpointAlgorithm;
}
public void setSslEndpointAlgorithm(String sslEndpointAlgorithm) {
this.sslEndpointAlgorithm = sslEndpointAlgorithm;
}
public String getSslKeymanagerAlgorithm() {
return sslKeymanagerAlgorithm;
}
public void setSslKeymanagerAlgorithm(String sslKeymanagerAlgorithm) {
this.sslKeymanagerAlgorithm = sslKeymanagerAlgorithm;
}
public String getSslKeyPassword() {
return sslKeyPassword;
}
public void setSslKeyPassword(String sslKeyPassword) {
this.sslKeyPassword = sslKeyPassword;
}
public String getSslKeystoreLocation() {
return sslKeystoreLocation;
}
public void setSslKeystoreLocation(String sslKeystoreLocation) {
this.sslKeystoreLocation = sslKeystoreLocation;
}
public String getSslKeystorePassword() {
return sslKeystorePassword;
}
public void setSslKeystorePassword(String sslKeystorePassword) {
this.sslKeystorePassword = sslKeystorePassword;
}
public String getSslKeystoreType() {
return sslKeystoreType;
}
public void setSslKeystoreType(String sslKeystoreType) {
this.sslKeystoreType = sslKeystoreType;
}
public String getSslProtocol() {
return sslProtocol;
}
public void setSslProtocol(String sslProtocol) {
this.sslProtocol = sslProtocol;
}
public String getSslProvider() {
return sslProvider;
}
public void setSslProvider(String sslProvider) {
this.sslProvider = sslProvider;
}
public String getSslTrustmanagerAlgorithm() {
return sslTrustmanagerAlgorithm;
}
public void setSslTrustmanagerAlgorithm(String sslTrustmanagerAlgorithm) {
this.sslTrustmanagerAlgorithm = sslTrustmanagerAlgorithm;
}
public String getSslTruststoreLocation() {
return sslTruststoreLocation;
}
public void setSslTruststoreLocation(String sslTruststoreLocation) {
this.sslTruststoreLocation = sslTruststoreLocation;
}
public String getSslTruststorePassword() {
return sslTruststorePassword;
}
public void setSslTruststorePassword(String sslTruststorePassword) {
this.sslTruststorePassword = sslTruststorePassword;
}
public String getSslTruststoreType() {
return sslTruststoreType;
}
public void setSslTruststoreType(String sslTruststoreType) {
this.sslTruststoreType = sslTruststoreType;
}
public Boolean getUseGlobalSslContextParameters() {
return useGlobalSslContextParameters;
}
public void setUseGlobalSslContextParameters(
Boolean useGlobalSslContextParameters) {
this.useGlobalSslContextParameters = useGlobalSslContextParameters;
}
}