Kafka alternatives and similar packages
Based on the "Big Data" category.
Alternatively, view Apache Kafka alternatives based on common mentions on social networks and blogs.
-
Deeplearning4J
Suite of tools for deploying and training deep learning models using the JVM. Highlights include model import for keras, tensorflow, and onnx/pytorch, a modular and tiny c++ library for running math code and a java based math library on top of the core c++ library. Also includes samediff: a pytorch/tensorflow like library for running deep learn... -
Reactive-kafka
Alpakka Kafka connector - Alpakka is a Reactive Enterprise Integration library for Java and Scala, based on Reactive Streams and Akka. -
Schemer
Schema registry for CSV, TSV, JSON, AVRO and Parquet schema. Supports schema inference and GraphQL API. -
GridScale
Scala library for accessing various file, batch systems, job schedulers and grid middlewares. -
Spark Utils
Basic framework utilities to quickly start writing production ready Apache Spark applications
SaaSHub - Software Alternatives and Reviews
* Code Quality Rankings and insights are calculated and provided by Lumnify.
They vary from L1 to L5 with "L5" being the highest.
Do you think we are missing an alternative of Kafka or a related project?
Popular Comparisons
README
Apache Kafka
See our web site for details on the project.
You need to have Java installed.
We build and test Apache Kafka with Java 8, 11 and 17. We set the release
parameter in javac and scalac
to 8
to ensure the generated binaries are compatible with Java 8 or higher (independently of the Java version
used for compilation). Java 8 support has been deprecated since Apache Kafka 3.0 and will be removed in Apache
Kafka 4.0 (see KIP-750 for more details).
Scala 2.12 and 2.13 are supported and 2.13 is used by default. Scala 2.12 support has been deprecated since Apache Kafka 3.0 and will be removed in Apache Kafka 4.0 (see KIP-751 for more details). See below for how to use a specific Scala version or all of the supported Scala versions.
Build a jar and run it
./gradlew jar
Follow instructions in https://kafka.apache.org/quickstart
Build source jar
./gradlew srcJar
Build aggregated javadoc
./gradlew aggregatedJavadoc
Build javadoc and scaladoc
./gradlew javadoc
./gradlew javadocJar # builds a javadoc jar for each module
./gradlew scaladoc
./gradlew scaladocJar # builds a scaladoc jar for each module
./gradlew docsJar # builds both (if applicable) javadoc and scaladoc jars for each module
Run unit/integration tests
./gradlew test # runs both unit and integration tests
./gradlew unitTest
./gradlew integrationTest
Force re-running tests without code change
./gradlew -Prerun-tests test
./gradlew -Prerun-tests unitTest
./gradlew -Prerun-tests integrationTest
Running a particular unit/integration test
./gradlew clients:test --tests RequestResponseTest
Repeatedly running a particular unit/integration test
I=0; while ./gradlew clients:test -Prerun-tests --tests RequestResponseTest --fail-fast; do (( I=$I+1 )); echo "Completed run: $I"; sleep 1; done
Running a particular test method within a unit/integration test
./gradlew core:test --tests kafka.api.ProducerFailureHandlingTest.testCannotSendToInternalTopic
./gradlew clients:test --tests org.apache.kafka.clients.MetadataTest.testTimeToNextUpdate
Running a particular unit/integration test with log4j output
Change the log4j setting in either clients/src/test/resources/log4j.properties
or core/src/test/resources/log4j.properties
./gradlew clients:test --tests RequestResponseTest
Specifying test retries
By default, each failed test is retried once up to a maximum of five retries per test run. Tests are retried at the end of the test task. Adjust these parameters in the following way:
./gradlew test -PmaxTestRetries=1 -PmaxTestRetryFailures=5
See Test Retry Gradle Plugin for more details.
Generating test coverage reports
Generate coverage reports for the whole project:
./gradlew reportCoverage -PenableTestCoverage=true -Dorg.gradle.parallel=false
Generate coverage for a single module, i.e.:
./gradlew clients:reportCoverage -PenableTestCoverage=true -Dorg.gradle.parallel=false
Building a binary release gzipped tar ball
./gradlew clean releaseTarGz
The release file can be found inside ./core/build/distributions/
.
Building auto generated messages
Sometimes it is only necessary to rebuild the RPC auto-generated message data when switching between branches, as they could fail due to code changes. You can just run:
./gradlew processMessages processTestMessages
Running a Kafka broker in KRaft mode
KAFKA_CLUSTER_ID="$(./bin/kafka-storage.sh random-uuid)"
./bin/kafka-storage.sh format -t $KAFKA_CLUSTER_ID -c config/kraft/server.properties
./bin/kafka-server-start.sh config/kraft/server.properties
Running a Kafka broker in ZooKeeper mode
./bin/zookeeper-server-start.sh config/zookeeper.properties
./bin/kafka-server-start.sh config/server.properties
Cleaning the build
./gradlew clean
Running a task with one of the Scala versions available (2.12.x or 2.13.x)
Note that if building the jars with a version other than 2.13.x, you need to set the SCALA_VERSION
variable or change it in bin/kafka-run-class.sh
to run the quick start.
You can pass either the major version (eg 2.12) or the full version (eg 2.12.7):
./gradlew -PscalaVersion=2.12 jar
./gradlew -PscalaVersion=2.12 test
./gradlew -PscalaVersion=2.12 releaseTarGz
Running a task with all the scala versions enabled by default
Invoke the gradlewAll
script followed by the task(s):
./gradlewAll test
./gradlewAll jar
./gradlewAll releaseTarGz
Running a task for a specific project
This is for core
, examples
and clients
./gradlew core:jar
./gradlew core:test
Streams has multiple sub-projects, but you can run all the tests:
./gradlew :streams:testAll
Listing all gradle tasks
./gradlew tasks
Building IDE project
Note that this is not strictly necessary (IntelliJ IDEA has good built-in support for Gradle projects, for example).
./gradlew eclipse
./gradlew idea
The eclipse
task has been configured to use ${project_dir}/build_eclipse
as Eclipse's build directory. Eclipse's default
build directory (${project_dir}/bin
) clashes with Kafka's scripts directory and we don't use Gradle's build directory
to avoid known issues with this configuration.
Publishing the jar for all versions of Scala and for all projects to maven
The recommended command is:
./gradlewAll publish
For backwards compatibility, the following also works:
./gradlewAll uploadArchives
Please note for this to work you should create/update ${GRADLE_USER_HOME}/gradle.properties
(typically, ~/.gradle/gradle.properties
) and assign the following variables
mavenUrl=
mavenUsername=
mavenPassword=
signing.keyId=
signing.password=
signing.secretKeyRingFile=
Publishing the streams quickstart archetype artifact to maven
For the Streams archetype project, one cannot use gradle to upload to maven; instead the mvn deploy
command needs to be called at the quickstart folder:
cd streams/quickstart
mvn deploy
Please note for this to work you should create/update user maven settings (typically, ${USER_HOME}/.m2/settings.xml
) to assign the following variables
<settings xmlns="http://maven.apache.org/SETTINGS/1.0.0"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="http://maven.apache.org/SETTINGS/1.0.0
https://maven.apache.org/xsd/settings-1.0.0.xsd">
...
<servers>
...
<server>
<id>apache.snapshots.https</id>
<username>${maven_username}</username>
<password>${maven_password}</password>
</server>
<server>
<id>apache.releases.https</id>
<username>${maven_username}</username>
<password>${maven_password}</password>
</server>
...
</servers>
...
Installing ALL the jars to the local Maven repository
The recommended command to build for both Scala 2.12 and 2.13 is:
./gradlewAll publishToMavenLocal
For backwards compatibility, the following also works:
./gradlewAll install
Installing specific projects to the local Maven repository
./gradlew -PskipSigning=true :streams:publishToMavenLocal
If needed, you can specify the Scala version with -PscalaVersion=2.13
.
Building the test jar
./gradlew testJar
Running code quality checks
There are two code quality analysis tools that we regularly run, spotbugs and checkstyle.
Checkstyle
Checkstyle enforces a consistent coding style in Kafka. You can run checkstyle using:
./gradlew checkstyleMain checkstyleTest
The checkstyle warnings will be found in reports/checkstyle/reports/main.html
and reports/checkstyle/reports/test.html
files in the
subproject build directories. They are also printed to the console. The build will fail if Checkstyle fails.
Spotbugs
Spotbugs uses static analysis to look for bugs in the code. You can run spotbugs using:
./gradlew spotbugsMain spotbugsTest -x test
The spotbugs warnings will be found in reports/spotbugs/main.html
and reports/spotbugs/test.html
files in the subproject build
directories. Use -PxmlSpotBugsReport=true to generate an XML report instead of an HTML one.
JMH microbenchmarks
We use JMH to write microbenchmarks that produce reliable results in the JVM.
See jmh-benchmarks/README.md for details on how to run the microbenchmarks.
Dependency Analysis
The gradle dependency debugging documentation mentions using the dependencies
or dependencyInsight
tasks to debug dependencies for the root project or individual subprojects.
Alternatively, use the allDeps
or allDepInsight
tasks for recursively iterating through all subprojects:
./gradlew allDeps
./gradlew allDepInsight --configuration runtimeClasspath --dependency com.fasterxml.jackson.core:jackson-databind
These take the same arguments as the builtin variants.
Determining if any dependencies could be updated
./gradlew dependencyUpdates
Common build options
The following options should be set with a -P
switch, for example ./gradlew -PmaxParallelForks=1 test
.
commitId
: sets the build commit ID as .git/HEAD might not be correct if there are local commits added for build purposes.mavenUrl
: sets the URL of the maven deployment repository (file://path/to/repo
can be used to point to a local repository).maxParallelForks
: maximum number of test processes to start in parallel. Defaults to the number of processors available to the JVM.maxScalacThreads
: maximum number of worker threads for the scalac backend. Defaults to the lowest of8
and the number of processors available to the JVM. The value must be between 1 and 16 (inclusive).ignoreFailures
: ignore test failures from junitshowStandardStreams
: shows standard out and standard error of the test JVM(s) on the console.skipSigning
: skips signing of artifacts.testLoggingEvents
: unit test events to be logged, separated by comma. For example./gradlew -PtestLoggingEvents=started,passed,skipped,failed test
.xmlSpotBugsReport
: enable XML reports for spotBugs. This also disables HTML reports as only one can be enabled at a time.maxTestRetries
: maximum number of retries for a failing test case.maxTestRetryFailures
: maximum number of test failures before retrying is disabled for subsequent tests.enableTestCoverage
: enables test coverage plugins and tasks, including bytecode enhancement of classes required to track said coverage. Note that this introduces some overhead when running tests and hence why it's disabled by default (the overhead varies, but 15-20% is a reasonable estimate).scalaOptimizerMode
: configures the optimizing behavior of the scala compiler, the value should be one ofnone
,method
,inline-kafka
orinline-scala
(the default isinline-kafka
).none
is the scala compiler default, which only eliminates unreachable code.method
also includes method-local optimizations.inline-kafka
adds inlining of methods within the kafka packages. Finally,inline-scala
also includes inlining of methods within the scala library (which avoids lambda allocations for methods likeOption.exists
).inline-scala
is only safe if the Scala library version is the same at compile time and runtime. Since we cannot guarantee this for all cases (for example, users may depend on the kafka jar for integration tests where they may include a scala library with a different version), we don't enable it by default. See https://www.lightbend.com/blog/scala-inliner-optimizer for more details.
Running system tests
See [tests/README.md](tests/README.md).
Running in Vagrant
See [vagrant/README.md](vagrant/README.md).
Contribution
Apache Kafka is interested in building the community; we would welcome any thoughts or patches. You can reach us on the Apache mailing lists.
To contribute follow the instructions here: