See our web site for details on the project.
You need to have Gradle and Java installed.
Kafka requires Gradle 5.0 or higher.
Java 8 should be used for building in order to support both Java 8 and Java 11 at runtime.
Scala 2.12 is used by default, see below for how to use a different Scala version or all of the supported Scala versions.
cd kafka_source_dir
gradle
Now everything else will work.
./gradlew jar
Follow instructions in https://kafka.apache.org/documentation.html#quickstart
./gradlew srcJar
./gradlew aggregatedJavadoc
./gradlew javadoc
./gradlew javadocJar # builds a javadoc jar for each module
./gradlew scaladoc
./gradlew scaladocJar # builds a scaladoc jar for each module
./gradlew docsJar # builds both (if applicable) javadoc and scaladoc jars for each module
./gradlew test # runs both unit and integration tests
./gradlew unitTest
./gradlew integrationTest
./gradlew cleanTest test
./gradlew cleanTest unitTest
./gradlew cleanTest integrationTest
./gradlew clients:test --tests RequestResponseTest
./gradlew core:test --tests kafka.api.ProducerFailureHandlingTest.testCannotSendToInternalTopic
./gradlew clients:test --tests org.apache.kafka.clients.MetadataTest.testMetadataUpdateWaitTime
Change the log4j setting in either clients/src/test/resources/log4j.properties
or core/src/test/resources/log4j.properties
./gradlew clients:test --tests RequestResponseTest
Generate coverage reports for the whole project:
./gradlew reportCoverage
Generate coverage for a single module, i.e.:
./gradlew clients:reportCoverage
./gradlew clean releaseTarGz
The above command will fail if you haven't set up the signing key. To bypass signing the artifact, you can run:
./gradlew clean releaseTarGz -x signArchives
The release file can be found inside ./core/build/distributions/
.
./gradlew clean
Note that if building the jars with a version other than 2.12.x, you need to set the SCALA_VERSION
variable or change it in bin/kafka-run-class.sh
to run the quick start.
You can pass either the major version (eg 2.12) or the full version (eg 2.12.7):
./gradlew -PscalaVersion=2.12 jar
./gradlew -PscalaVersion=2.12 test
./gradlew -PscalaVersion=2.12 releaseTarGz
Append All
to the task name:
./gradlew testAll
./gradlew jarAll
./gradlew releaseTarGzAll
This is for core
, examples
and clients
./gradlew core:jar
./gradlew core:test
./gradlew tasks
Note that this is not strictly necessary (IntelliJ IDEA has good built-in support for Gradle projects, for example).
./gradlew eclipse
./gradlew idea
The eclipse
task has been configured to use ${project_dir}/build_eclipse
as Eclipse's build directory. Eclipse's default
build directory (${project_dir}/bin
) clashes with Kafka's scripts directory and we don't use Gradle's build directory
to avoid known issues with this configuration.
./gradlew uploadArchivesAll
Please note for this to work you should create/update ${GRADLE_USER_HOME}/gradle.properties
(typically, ~/.gradle/gradle.properties
) and assign the following variables
mavenUrl=
mavenUsername=
mavenPassword=
signing.keyId=
signing.password=
signing.secretKeyRingFile=
For the Streams archetype project, one cannot use gradle to upload to maven; instead the mvn deploy
command needs to be called at the quickstart folder:
cd streams/quickstart
mvn deploy
Please note for this to work you should create/update user maven settings (typically, ${USER_HOME}/.m2/settings.xml
) to assign the following variables
<settings xmlns="http://maven.apache.org/SETTINGS/1.0.0"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="http://maven.apache.org/SETTINGS/1.0.0
https://maven.apache.org/xsd/settings-1.0.0.xsd">
...
<servers>
...
<server>
<id>apache.snapshots.https</id>
<username>${maven_username}</username>
<password>${maven_password}</password>
</server>
<server>
<id>apache.releases.https</id>
<username>${maven_username}</username>
<password>${maven_password}</password>
</server>
...
</servers>
...
./gradlew installAll
./gradlew testJar
./gradlew core:dependencies --configuration runtime
./gradlew dependencyUpdates
There are two code quality analysis tools that we regularly run, spotbugs and checkstyle.
Checkstyle enforces a consistent coding style in Kafka. You can run checkstyle using:
./gradlew checkstyleMain checkstyleTest
The checkstyle warnings will be found in reports/checkstyle/reports/main.html
and reports/checkstyle/reports/test.html
files in the
subproject build directories. They are also are printed to the console. The build will fail if Checkstyle fails.
Spotbugs uses static analysis to look for bugs in the code. You can run spotbugs using:
./gradlew spotbugsMain spotbugsTest -x test
The spotbugs warnings will be found in reports/spotbugs/main.html
and reports/spotbugs/test.html
files in the subproject build
directories. Use -PxmlSpotBugsReport=true to generate an XML report instead of an HTML one.
The following options should be set with a -P
switch, for example ./gradlew -PmaxParallelForks=1 test
.
commitId
: sets the build commit ID as .git/HEAD might not be correct if there are local commits added for build purposes.mavenUrl
: sets the URL of the maven deployment repository (file://path/to/repo
can be used to point to a local repository).maxParallelForks
: limits the maximum number of processes for each task.showStandardStreams
: shows standard out and standard error of the test JVM(s) on the console.skipSigning
: skips signing of artifacts.testLoggingEvents
: unit test events to be logged, separated by comma. For example./gradlew -PtestLoggingEvents=started,passed,skipped,failed test
.xmlSpotBugsReport
: enable XML reports for spotBugs. This also disables HTML reports as only one can be enabled at a time.
See vagrant/README.md.
Apache Kafka is interested in building the community; we would welcome any thoughts or patches. You can reach us on the Apache mailing lists.
To contribute follow the instructions here:
- Kafka源码分析-网络层-1
- Kafka源码分析-网络层-2
- Kafka源码分析-网络层-3
- Kafka源码分析-配置文件
- Kafka源码分析-启动流程
- Kafka基础组件和辅助类库简介
- Kafka的Request和Response
- Kafka初始化流程与请求处理
- KafkaController分析1-选主和Failover
- KafkaController分析2-NetworkClient分析
- KafkaController分析3-ControllerChannelManager
- KafkaController分析4-Partition选主
- KafkaController分析5-Partition状态机
- KafkaController分析6-Replica状态机
- KafkaController分析7-启动流程
- Kafka集群建立过程分析
- Kafka集群Metadata管理
- KafkaController分析8-broker挂掉
- ReplicaManager源码解析1-相关基础类库
- ReplicaManager源码解析2-LeaderAndIsr 请求响应
- Kafka中的时间轮
- DelayedOperationPurgatory--谜之炼狱
- Kafka的日志管理模块--LogManager
- Kafka中Message存储相关类大揭密
- Kafka消息的磁盘存储
- Kafka中的消息操作的层级调用关系
- Kafka是如何处理客户端发送的数据的?
- ISR列表是如何变化的?
- Kafka的消息是如何被消费的?
- Kafka的日志清理-LogCleane
- Kafka运维填坑
- Kafka重置消费的Offset
- 记一次Kafka集群的故障恢复
- 使用多数据中心部署来应对灾难恢复(一)
- 使用多数据中心部署来应对灾难恢复(二)
- 优化你的Apache Kafka部署