blob: 9729341f6a114c1c66708eb675e8bf8dfdd103f1 [file] [log] [blame]
// kafka-connector options: START
[[camel-debezium-mongodb-kafka-connector-source]]
= camel-debezium-mongodb-kafka-connector source configuration
When using camel-debezium-mongodb-kafka-connector as source make sure to use the following Maven dependency to have support for the connector:
[source,xml]
----
<dependency>
<groupId>org.apache.camel.kafkaconnector</groupId>
<artifactId>camel-debezium-mongodb-kafka-connector</artifactId>
<version>x.x.x</version>
<!-- use the same version as your Camel Kafka connector version -->
</dependency>
----
The camel-debezium-mongodb source connector supports 48 options, which are listed below.
[width="100%",cols="2,5,^1,2",options="header"]
|===
| Name | Description | Default | Priority
| *camel.source.path.name* | Unique name for the connector. Attempting to register again with the same name will fail. | null | HIGH
| *camel.source.endpoint.additionalProperties* | Additional properties for debezium components in case they can't be set directly on the camel configurations (e.g: setting Kafka Connect properties needed by Debezium engine, for example setting KafkaOffsetBackingStore), the properties have to be prefixed with additionalProperties.. E.g: additionalProperties.transactional.id=12345&additionalProperties.schema.registry.url=\http://localhost:8811/avro | null | MEDIUM
| *camel.source.endpoint.bridgeErrorHandler* | Allows for bridging the consumer to the Camel routing Error Handler, which mean any exceptions occurred while the consumer is trying to pickup incoming messages, or the likes, will now be processed as a message and handled by the routing Error Handler. By default the consumer will use the org.apache.camel.spi.ExceptionHandler to deal with exceptions, that will be logged at WARN or ERROR level and ignored. | false | MEDIUM
| *camel.source.endpoint.internalKeyConverter* | The Converter class that should be used to serialize and deserialize key data for offsets. The default is JSON converter. | "org.apache.kafka.connect.json.JsonConverter" | MEDIUM
| *camel.source.endpoint.internalValueConverter* | The Converter class that should be used to serialize and deserialize value data for offsets. The default is JSON converter. | "org.apache.kafka.connect.json.JsonConverter" | MEDIUM
| *camel.source.endpoint.offsetCommitPolicy* | The name of the Java class of the commit policy. It defines when offsets commit has to be triggered based on the number of events processed and the time elapsed since the last commit. This class must implement the interface 'OffsetCommitPolicy'. The default is a periodic commit policy based upon time intervals. | "io.debezium.embedded.spi.OffsetCommitPolicy.PeriodicCommitOffsetPolicy" | MEDIUM
| *camel.source.endpoint.offsetCommitTimeoutMs* | Maximum number of milliseconds to wait for records to flush and partition offset data to be committed to offset storage before cancelling the process and restoring the offset data to be committed in a future attempt. The default is 5 seconds. | 5000L | MEDIUM
| *camel.source.endpoint.offsetFlushIntervalMs* | Interval at which to try committing offsets. The default is 1 minute. | 60000L | MEDIUM
| *camel.source.endpoint.offsetStorage* | The name of the Java class that is responsible for persistence of connector offsets. | "org.apache.kafka.connect.storage.FileOffsetBackingStore" | MEDIUM
| *camel.source.endpoint.offsetStorageFileName* | Path to file where offsets are to be stored. Required when offset.storage is set to the FileOffsetBackingStore. | null | MEDIUM
| *camel.source.endpoint.offsetStoragePartitions* | The number of partitions used when creating the offset storage topic. Required when offset.storage is set to the 'KafkaOffsetBackingStore'. | null | MEDIUM
| *camel.source.endpoint.offsetStorageReplication Factor* | Replication factor used when creating the offset storage topic. Required when offset.storage is set to the KafkaOffsetBackingStore | null | MEDIUM
| *camel.source.endpoint.offsetStorageTopic* | The name of the Kafka topic where offsets are to be stored. Required when offset.storage is set to the KafkaOffsetBackingStore. | null | MEDIUM
| *camel.source.endpoint.exceptionHandler* | To let the consumer use a custom ExceptionHandler. Notice if the option bridgeErrorHandler is enabled then this option is not in use. By default the consumer will deal with exceptions, that will be logged at WARN or ERROR level and ignored. | null | MEDIUM
| *camel.source.endpoint.exchangePattern* | Sets the exchange pattern when the consumer creates an exchange. One of: [InOnly] [InOut] [InOptionalOut] | null | MEDIUM
| *camel.source.endpoint.basicPropertyBinding* | Whether the endpoint should use basic property binding (Camel 2.x) or the newer property binding with additional capabilities | false | MEDIUM
| *camel.source.endpoint.synchronous* | Sets whether synchronous processing should be strictly used, or Camel is allowed to use asynchronous processing (if supported). | false | MEDIUM
| *camel.source.endpoint.collectionBlacklist* | Description is not available here, please check Debezium website for corresponding key 'collection.blacklist' description. | null | MEDIUM
| *camel.source.endpoint.collectionWhitelist* | The collections for which changes are to be captured | null | MEDIUM
| *camel.source.endpoint.connectBackoffInitialDelayMs* | The initial delay when trying to reconnect to a primary after a connection cannot be made or when no primary is available. Defaults to 1 second (1000 ms). | 1000L | MEDIUM
| *camel.source.endpoint.connectBackoffMaxDelayMs* | The maximum delay when trying to reconnect to a primary after a connection cannot be made or when no primary is available. Defaults to 120 second (120,000 ms). | 120000L | MEDIUM
| *camel.source.endpoint.connectMaxAttempts* | Maximum number of failed connection attempts to a replica set primary before an exception occurs and task is aborted. Defaults to 16, which with the defaults for 'connect.backoff.initial.delay.ms' and 'connect.backoff.max.delay.ms' results in just over 20 minutes of attempts before failing. | 16 | MEDIUM
| *camel.source.endpoint.databaseBlacklist* | The databases for which changes are to be excluded | null | MEDIUM
| *camel.source.endpoint.databaseHistoryFileFilename* | The path to the file that will be used to record the database history | null | MEDIUM
| *camel.source.endpoint.databaseWhitelist* | The databases for which changes are to be captured | null | MEDIUM
| *camel.source.endpoint.fieldBlacklist* | Description is not available here, please check Debezium website for corresponding key 'field.blacklist' description. | null | MEDIUM
| *camel.source.endpoint.fieldRenames* | Description is not available here, please check Debezium website for corresponding key 'field.renames' description. | null | MEDIUM
| *camel.source.endpoint.heartbeatIntervalMs* | Length of an interval in milli-seconds in in which the connector periodically sends heartbeat messages to a heartbeat topic. Use 0 to disable heartbeat messages. Disabled by default. | 0 | MEDIUM
| *camel.source.endpoint.heartbeatTopicsPrefix* | The prefix that is used to name heartbeat topics.Defaults to __debezium-heartbeat. | "__debezium-heartbeat" | MEDIUM
| *camel.source.endpoint.initialSyncMaxThreads* | Maximum number of threads used to perform an intial sync of the collections in a replica set. Defaults to 1. | 1 | MEDIUM
| *camel.source.endpoint.maxBatchSize* | Maximum size of each batch of source records. Defaults to 2048. | 2048 | MEDIUM
| *camel.source.endpoint.maxQueueSize* | Maximum size of the queue for change events read from the database log but not yet recorded or forwarded. Defaults to 8192, and should always be larger than the maximum batch size. | 8192 | MEDIUM
| *camel.source.endpoint.mongodbHosts* | The hostname and port pairs (in the form 'host' or 'host:port') of the MongoDB server(s) in the replica set. | null | MEDIUM
| *camel.source.endpoint.mongodbMembersAutoDiscover* | Specifies whether the addresses in 'hosts' are seeds that should be used to discover all members of the cluster or replica set ('true'), or whether the address(es) in 'hosts' should be used as is ('false'). The default is 'true'. | true | MEDIUM
| *camel.source.endpoint.mongodbName* | Unique name that identifies the MongoDB replica set or cluster and all recorded offsets, andthat is used as a prefix for all schemas and topics. Each distinct MongoDB installation should have a separate namespace and monitored by at most one Debezium connector. | null | HIGH
| *camel.source.endpoint.mongodbPassword* | Password to be used when connecting to MongoDB, if necessary. | null | HIGH
| *camel.source.endpoint.mongodbSslEnabled* | Should connector use SSL to connect to MongoDB instances | false | MEDIUM
| *camel.source.endpoint.mongodbSslInvalidHostname Allowed* | Whether invalid host names are allowed when using SSL. If true the connection will not prevent man-in-the-middle attacks | false | MEDIUM
| *camel.source.endpoint.mongodbUser* | Database user for connecting to MongoDB, if necessary. | null | MEDIUM
| *camel.source.endpoint.pollIntervalMs* | Frequency in milliseconds to wait for new change events to appear after receiving no events. Defaults to 500ms. | 500L | MEDIUM
| *camel.source.endpoint.snapshotDelayMs* | The number of milliseconds to delay before a snapshot will begin. | 0L | MEDIUM
| *camel.source.endpoint.snapshotFetchSize* | The maximum number of records that should be loaded into memory while performing a snapshot | null | MEDIUM
| *camel.source.endpoint.snapshotMode* | The criteria for running a snapshot upon startup of the connector. Options include: 'initial' (the default) to specify the connector should always perform an initial sync when required; 'never' to specify the connector should never perform an initial sync | "initial" | MEDIUM
| *camel.source.endpoint.sourceStructVersion* | A version of the format of the publicly visible source part in the message | "v2" | MEDIUM
| *camel.source.endpoint.tombstonesOnDelete* | Whether delete operations should be represented by a delete event and a subsquenttombstone event (true) or only by a delete event (false). Emitting the tombstone event (the default behavior) allows Kafka to completely delete all events pertaining to the given key once the source record got deleted. | false | MEDIUM
| *camel.component.debezium-mongodb.bridgeError Handler* | Allows for bridging the consumer to the Camel routing Error Handler, which mean any exceptions occurred while the consumer is trying to pickup incoming messages, or the likes, will now be processed as a message and handled by the routing Error Handler. By default the consumer will use the org.apache.camel.spi.ExceptionHandler to deal with exceptions, that will be logged at WARN or ERROR level and ignored. | false | MEDIUM
| *camel.component.debezium-mongodb.configuration* | Allow pre-configured Configurations to be set. | null | MEDIUM
| *camel.component.debezium-mongodb.basicProperty Binding* | Whether the component should use basic property binding (Camel 2.x) or the newer property binding with additional capabilities | false | MEDIUM
|===
// kafka-connector options: END