| Apache Kafka |
| ================= |
| See our [web site](https://kafka.apache.org) for details on the project. |
| |
| You need to have [Java](http://www.oracle.com/technetwork/java/javase/downloads/index.html) installed. |
| |
| We build and test Apache Kafka with Java 8, 11, 17 and 21. We set the `release` parameter in javac and scalac |
| to `8` to ensure the generated binaries are compatible with Java 8 or higher (independently of the Java version |
| used for compilation). Java 8 support project-wide has been deprecated since Apache Kafka 3.0, Java 11 support for |
| the broker and tools has been deprecated since Apache Kafka 3.7 and removal of both is planned for Apache Kafka 4.0 ( |
| see [KIP-750](https://cwiki.apache.org/confluence/pages/viewpage.action?pageId=181308223) and |
| [KIP-1013](https://cwiki.apache.org/confluence/pages/viewpage.action?pageId=284789510) for more details). |
| |
| Scala 2.12 and 2.13 are supported and 2.13 is used by default. Scala 2.12 support has been deprecated since |
| Apache Kafka 3.0 and will be removed in Apache Kafka 4.0 (see [KIP-751](https://cwiki.apache.org/confluence/pages/viewpage.action?pageId=181308218) |
| for more details). See below for how to use a specific Scala version or all of the supported Scala versions. |
| |
| ### Build a jar and run it ### |
| ./gradlew jar |
| |
| Follow instructions in https://kafka.apache.org/quickstart |
| |
| ### Build source jar ### |
| ./gradlew srcJar |
| |
| ### Build aggregated javadoc ### |
| ./gradlew aggregatedJavadoc |
| |
| ### Build javadoc and scaladoc ### |
| ./gradlew javadoc |
| ./gradlew javadocJar # builds a javadoc jar for each module |
| ./gradlew scaladoc |
| ./gradlew scaladocJar # builds a scaladoc jar for each module |
| ./gradlew docsJar # builds both (if applicable) javadoc and scaladoc jars for each module |
| |
| ### Run unit/integration tests ### |
| ./gradlew test # runs both unit and integration tests |
| ./gradlew unitTest |
| ./gradlew integrationTest |
| |
| ### Force re-running tests without code change ### |
| ./gradlew test --rerun |
| ./gradlew unitTest --rerun |
| ./gradlew integrationTest --rerun |
| |
| ### Running a particular unit/integration test ### |
| ./gradlew clients:test --tests RequestResponseTest |
| |
| ### Repeatedly running a particular unit/integration test ### |
| I=0; while ./gradlew clients:test --tests RequestResponseTest --rerun --fail-fast; do (( I=$I+1 )); echo "Completed run: $I"; sleep 1; done |
| |
| ### Running a particular test method within a unit/integration test ### |
| ./gradlew core:test --tests kafka.api.ProducerFailureHandlingTest.testCannotSendToInternalTopic |
| ./gradlew clients:test --tests org.apache.kafka.clients.MetadataTest.testTimeToNextUpdate |
| |
| ### Running a particular unit/integration test with log4j output ### |
| Change the log4j setting in either `clients/src/test/resources/log4j.properties` or `core/src/test/resources/log4j.properties` |
| |
| ./gradlew clients:test --tests RequestResponseTest |
| |
| ### Specifying test retries ### |
| By default, each failed test is retried once up to a maximum of five retries per test run. Tests are retried at the end of the test task. Adjust these parameters in the following way: |
| |
| ./gradlew test -PmaxTestRetries=1 -PmaxTestRetryFailures=5 |
| |
| See [Test Retry Gradle Plugin](https://github.com/gradle/test-retry-gradle-plugin) for more details. |
| |
| ### Generating test coverage reports ### |
| Generate coverage reports for the whole project: |
| |
| ./gradlew reportCoverage -PenableTestCoverage=true -Dorg.gradle.parallel=false |
| |
| Generate coverage for a single module, i.e.: |
| |
| ./gradlew clients:reportCoverage -PenableTestCoverage=true -Dorg.gradle.parallel=false |
| |
| ### Building a binary release gzipped tar ball ### |
| ./gradlew clean releaseTarGz |
| |
| The release file can be found inside `./core/build/distributions/`. |
| |
| ### Building auto generated messages ### |
| Sometimes it is only necessary to rebuild the RPC auto-generated message data when switching between branches, as they could |
| fail due to code changes. You can just run: |
| |
| ./gradlew processMessages processTestMessages |
| |
| ### Running a Kafka broker in KRaft mode |
| |
| KAFKA_CLUSTER_ID="$(./bin/kafka-storage.sh random-uuid)" |
| ./bin/kafka-storage.sh format -t $KAFKA_CLUSTER_ID -c config/kraft/server.properties |
| ./bin/kafka-server-start.sh config/kraft/server.properties |
| |
| ### Running a Kafka broker in ZooKeeper mode |
| |
| ./bin/zookeeper-server-start.sh config/zookeeper.properties |
| ./bin/kafka-server-start.sh config/server.properties |
| |
| ### Cleaning the build ### |
| ./gradlew clean |
| |
| ### Running a task with one of the Scala versions available (2.12.x or 2.13.x) ### |
| *Note that if building the jars with a version other than 2.13.x, you need to set the `SCALA_VERSION` variable or change it in `bin/kafka-run-class.sh` to run the quick start.* |
| |
| You can pass either the major version (eg 2.12) or the full version (eg 2.12.7): |
| |
| ./gradlew -PscalaVersion=2.12 jar |
| ./gradlew -PscalaVersion=2.12 test |
| ./gradlew -PscalaVersion=2.12 releaseTarGz |
| |
| ### Running a task with all the scala versions enabled by default ### |
| |
| Invoke the `gradlewAll` script followed by the task(s): |
| |
| ./gradlewAll test |
| ./gradlewAll jar |
| ./gradlewAll releaseTarGz |
| |
| ### Running a task for a specific project ### |
| This is for `core`, `examples` and `clients` |
| |
| ./gradlew core:jar |
| ./gradlew core:test |
| |
| Streams has multiple sub-projects, but you can run all the tests: |
| |
| ./gradlew :streams:testAll |
| |
| ### Listing all gradle tasks ### |
| ./gradlew tasks |
| |
| ### Building IDE project #### |
| *Note that this is not strictly necessary (IntelliJ IDEA has good built-in support for Gradle projects, for example).* |
| |
| ./gradlew eclipse |
| ./gradlew idea |
| |
| The `eclipse` task has been configured to use `${project_dir}/build_eclipse` as Eclipse's build directory. Eclipse's default |
| build directory (`${project_dir}/bin`) clashes with Kafka's scripts directory and we don't use Gradle's build directory |
| to avoid known issues with this configuration. |
| |
| ### Publishing the jar for all versions of Scala and for all projects to maven ### |
| The recommended command is: |
| |
| ./gradlewAll publish |
| |
| For backwards compatibility, the following also works: |
| |
| ./gradlewAll uploadArchives |
| |
| Please note for this to work you should create/update `${GRADLE_USER_HOME}/gradle.properties` (typically, `~/.gradle/gradle.properties`) and assign the following variables |
| |
| mavenUrl= |
| mavenUsername= |
| mavenPassword= |
| signing.keyId= |
| signing.password= |
| signing.secretKeyRingFile= |
| |
| ### Publishing the streams quickstart archetype artifact to maven ### |
| For the Streams archetype project, one cannot use gradle to upload to maven; instead the `mvn deploy` command needs to be called at the quickstart folder: |
| |
| cd streams/quickstart |
| mvn deploy |
| |
| Please note for this to work you should create/update user maven settings (typically, `${USER_HOME}/.m2/settings.xml`) to assign the following variables |
| |
| <settings xmlns="http://maven.apache.org/SETTINGS/1.0.0" |
| xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" |
| xsi:schemaLocation="http://maven.apache.org/SETTINGS/1.0.0 |
| https://maven.apache.org/xsd/settings-1.0.0.xsd"> |
| ... |
| <servers> |
| ... |
| <server> |
| <id>apache.snapshots.https</id> |
| <username>${maven_username}</username> |
| <password>${maven_password}</password> |
| </server> |
| <server> |
| <id>apache.releases.https</id> |
| <username>${maven_username}</username> |
| <password>${maven_password}</password> |
| </server> |
| ... |
| </servers> |
| ... |
| |
| |
| ### Installing ALL the jars to the local Maven repository ### |
| The recommended command to build for both Scala 2.12 and 2.13 is: |
| |
| ./gradlewAll publishToMavenLocal |
| |
| For backwards compatibility, the following also works: |
| |
| ./gradlewAll install |
| |
| ### Installing specific projects to the local Maven repository ### |
| |
| ./gradlew -PskipSigning=true :streams:publishToMavenLocal |
| |
| If needed, you can specify the Scala version with `-PscalaVersion=2.13`. |
| |
| ### Building the test jar ### |
| ./gradlew testJar |
| |
| ### Running code quality checks ### |
| There are two code quality analysis tools that we regularly run, spotbugs and checkstyle. |
| |
| #### Checkstyle #### |
| Checkstyle enforces a consistent coding style in Kafka. |
| You can run checkstyle using: |
| |
| ./gradlew checkstyleMain checkstyleTest |
| |
| The checkstyle warnings will be found in `reports/checkstyle/reports/main.html` and `reports/checkstyle/reports/test.html` files in the |
| subproject build directories. They are also printed to the console. The build will fail if Checkstyle fails. |
| |
| #### Spotbugs #### |
| Spotbugs uses static analysis to look for bugs in the code. |
| You can run spotbugs using: |
| |
| ./gradlew spotbugsMain spotbugsTest -x test |
| |
| The spotbugs warnings will be found in `reports/spotbugs/main.html` and `reports/spotbugs/test.html` files in the subproject build |
| directories. Use -PxmlSpotBugsReport=true to generate an XML report instead of an HTML one. |
| |
| ### JMH microbenchmarks ### |
| We use [JMH](https://openjdk.java.net/projects/code-tools/jmh/) to write microbenchmarks that produce reliable results in the JVM. |
| |
| See [jmh-benchmarks/README.md](https://github.com/apache/kafka/blob/trunk/jmh-benchmarks/README.md) for details on how to run the microbenchmarks. |
| |
| ### Dependency Analysis ### |
| |
| The gradle [dependency debugging documentation](https://docs.gradle.org/current/userguide/viewing_debugging_dependencies.html) mentions using the `dependencies` or `dependencyInsight` tasks to debug dependencies for the root project or individual subprojects. |
| |
| Alternatively, use the `allDeps` or `allDepInsight` tasks for recursively iterating through all subprojects: |
| |
| ./gradlew allDeps |
| |
| ./gradlew allDepInsight --configuration runtimeClasspath --dependency com.fasterxml.jackson.core:jackson-databind |
| |
| These take the same arguments as the builtin variants. |
| |
| ### Determining if any dependencies could be updated ### |
| ./gradlew dependencyUpdates |
| |
| ### Common build options ### |
| |
| The following options should be set with a `-P` switch, for example `./gradlew -PmaxParallelForks=1 test`. |
| |
| * `commitId`: sets the build commit ID as .git/HEAD might not be correct if there are local commits added for build purposes. |
| * `mavenUrl`: sets the URL of the maven deployment repository (`file://path/to/repo` can be used to point to a local repository). |
| * `maxParallelForks`: maximum number of test processes to start in parallel. Defaults to the number of processors available to the JVM. |
| * `maxScalacThreads`: maximum number of worker threads for the scalac backend. Defaults to the lowest of `8` and the number of processors |
| available to the JVM. The value must be between 1 and 16 (inclusive). |
| * `ignoreFailures`: ignore test failures from junit |
| * `showStandardStreams`: shows standard out and standard error of the test JVM(s) on the console. |
| * `skipSigning`: skips signing of artifacts. |
| * `testLoggingEvents`: unit test events to be logged, separated by comma. For example `./gradlew -PtestLoggingEvents=started,passed,skipped,failed test`. |
| * `xmlSpotBugsReport`: enable XML reports for spotBugs. This also disables HTML reports as only one can be enabled at a time. |
| * `maxTestRetries`: maximum number of retries for a failing test case. |
| * `maxTestRetryFailures`: maximum number of test failures before retrying is disabled for subsequent tests. |
| * `enableTestCoverage`: enables test coverage plugins and tasks, including bytecode enhancement of classes required to track said |
| coverage. Note that this introduces some overhead when running tests and hence why it's disabled by default (the overhead |
| varies, but 15-20% is a reasonable estimate). |
| * `keepAliveMode`: configures the keep alive mode for the Gradle compilation daemon - reuse improves start-up time. The values should |
| be one of `daemon` or `session` (the default is `daemon`). `daemon` keeps the daemon alive until it's explicitly stopped while |
| `session` keeps it alive until the end of the build session. This currently only affects the Scala compiler, see |
| https://github.com/gradle/gradle/pull/21034 for a PR that attempts to do the same for the Java compiler. |
| * `scalaOptimizerMode`: configures the optimizing behavior of the scala compiler, the value should be one of `none`, `method`, `inline-kafka` or |
| `inline-scala` (the default is `inline-kafka`). `none` is the scala compiler default, which only eliminates unreachable code. `method` also |
| includes method-local optimizations. `inline-kafka` adds inlining of methods within the kafka packages. Finally, `inline-scala` also |
| includes inlining of methods within the scala library (which avoids lambda allocations for methods like `Option.exists`). `inline-scala` is |
| only safe if the Scala library version is the same at compile time and runtime. Since we cannot guarantee this for all cases (for example, users |
| may depend on the kafka jar for integration tests where they may include a scala library with a different version), we don't enable it by |
| default. See https://www.lightbend.com/blog/scala-inliner-optimizer for more details. |
| |
| ### Running system tests ### |
| |
| See [tests/README.md](tests/README.md). |
| |
| ### Running in Vagrant ### |
| |
| See [vagrant/README.md](vagrant/README.md). |
| |
| ### Contribution ### |
| |
| Apache Kafka is interested in building the community; we would welcome any thoughts or [patches](https://issues.apache.org/jira/browse/KAFKA). You can reach us [on the Apache mailing lists](http://kafka.apache.org/contact.html). |
| |
| To contribute follow the instructions here: |
| * https://kafka.apache.org/contributing.html |