| ## |
| # Licensed to the Apache Software Foundation (ASF) under one or more |
| # contributor license agreements. See the NOTICE file distributed with |
| # this work for additional information regarding copyright ownership. |
| # The ASF licenses this file to You under the Apache License, Version 2.0 |
| # (the "License"); you may not use this file except in compliance with |
| # the License. You may obtain a copy of the License at |
| # |
| # http://www.apache.org/licenses/LICENSE-2.0 |
| # |
| # Unless required by applicable law or agreed to in writing, software |
| # distributed under the License is distributed on an "AS IS" BASIS, |
| # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| # See the License for the specific language governing permissions and |
| # limitations under the License. |
| ## |
| |
| # This file contains some of the configurations for the Kafka Connect distributed worker. This file is intended |
| # to be used with the examples, and some settings may differ from those used in a production system, especially |
| # the `bootstrap.servers` and those specifying replication factors. |
| |
| # A list of host/port pairs to use for establishing the initial connection to the Kafka cluster. |
| bootstrap.servers=localhost:9092 |
| |
| # unique name for the cluster, used in forming the Connect cluster group. Note that this must not conflict with consumer group IDs |
| group.id=connect-cluster |
| |
| # The converters specify the format of data in Kafka and how to translate it into Connect data. Every Connect user will |
| # need to configure these based on the format they want their data in when loaded from or stored into Kafka |
| key.converter=org.apache.kafka.connect.json.JsonConverter |
| value.converter=org.apache.kafka.connect.json.JsonConverter |
| # Converter-specific settings can be passed in by prefixing the Converter's setting with the converter we want to apply |
| # it to |
| key.converter.schemas.enable=true |
| value.converter.schemas.enable=true |
| |
| # Topic to use for storing offsets. This topic should have many partitions and be replicated and compacted. |
| # Kafka Connect will attempt to create the topic automatically when needed, but you can always manually create |
| # the topic before starting Kafka Connect if a specific topic configuration is needed. |
| # Most users will want to use the built-in default replication factor of 3 or in some cases even specify a larger value. |
| # Since this means there must be at least as many brokers as the maximum replication factor used, we'd like to be able |
| # to run this example on a single-broker cluster and so here we instead set the replication factor to 1. |
| offset.storage.topic=connect-offsets |
| offset.storage.replication.factor=1 |
| #offset.storage.partitions=25 |
| |
| # Topic to use for storing connector and task configurations; note that this should be a single partition, highly replicated, |
| # and compacted topic. Kafka Connect will attempt to create the topic automatically when needed, but you can always manually create |
| # the topic before starting Kafka Connect if a specific topic configuration is needed. |
| # Most users will want to use the built-in default replication factor of 3 or in some cases even specify a larger value. |
| # Since this means there must be at least as many brokers as the maximum replication factor used, we'd like to be able |
| # to run this example on a single-broker cluster and so here we instead set the replication factor to 1. |
| config.storage.topic=connect-configs |
| config.storage.replication.factor=1 |
| |
| # Topic to use for storing statuses. This topic can have multiple partitions and should be replicated and compacted. |
| # Kafka Connect will attempt to create the topic automatically when needed, but you can always manually create |
| # the topic before starting Kafka Connect if a specific topic configuration is needed. |
| # Most users will want to use the built-in default replication factor of 3 or in some cases even specify a larger value. |
| # Since this means there must be at least as many brokers as the maximum replication factor used, we'd like to be able |
| # to run this example on a single-broker cluster and so here we instead set the replication factor to 1. |
| status.storage.topic=connect-status |
| status.storage.replication.factor=1 |
| #status.storage.partitions=5 |
| |
| # Flush much faster than normal, which is useful for testing/debugging |
| offset.flush.interval.ms=10000 |
| |
| # These are provided to inform the user about the presence of the REST host and port configs |
| # Hostname & Port for the REST API to listen on. If this is set, it will bind to the interface used to listen to requests. |
| #rest.host.name= |
| #rest.port=8083 |
| |
| # The Hostname & Port that will be given out to other workers to connect to i.e. URLs that are routable from other servers. |
| #rest.advertised.host.name= |
| #rest.advertised.port= |
| |
| # Set to a list of filesystem paths separated by commas (,) to enable class loading isolation for plugins |
| # (connectors, converters, transformations). The list should consist of top level directories that include |
| # any combination of: |
| # a) directories immediately containing jars with plugins and their dependencies |
| # b) uber-jars with plugins and their dependencies |
| # c) directories immediately containing the package directory structure of classes of plugins and their dependencies |
| # Examples: |
| # plugin.path=/usr/local/share/java,/usr/local/share/kafka/plugins,/opt/connectors, |
| #plugin.path= |