Based on kafka-docker by Wurstmeister.
Updated and adapted for running properly on my Mac OS X.
Dockerfile for Apache Kafka
##Pre-Requisites
- install docker-compose https://docs.docker.com/compose/install/
- modify the
KAFKA_ADVERTISED_HOST_NAME
indocker-compose.yml
to match your docker host IP (Note: Do not use localhost or 127.0.0.1 as the host ip if you want to run multiple brokers.) (find this withdocker-compose ip default
; on a "fresh" Mac with docker 1.9 this typically resolves to192.168.99.100
) - if you want to customise any Kafka parameters, simply add them as environment variables in
docker-compose.yml
, e.g. in order to increase themessage.max.bytes
parameter set the environment toKAFKA_MESSAGE_MAX_BYTES: 2000000
. To turn off automatic topic creation setKAFKA_AUTO_CREATE_TOPICS_ENABLE: 'false'
##Usage
Start a cluster:
docker-compose up -d
(this will download approx. 1GB and take 5 minutes)
To validate correct operation:
- two topics should have been created:
$ kafka/bin/kafka-topics.sh --list --zookeeper 192.168.99.100:2181
Topic1
Topic2
- Kafka console producer and consumer work
In one terminal:
$ kafka/bin/kafka-console-producer.sh --topic Topic1 --broker-list 192.168.99.100:9092
[2015-11-17 08:41:34,751] WARN Property topic is not valid (kafka.utils.VerifiableProperties)
foo
bar
In another terminal:
$ kafka/bin/kafka-console-consumer.sh --topic Topic1 --zookeeper 192.168.99.100:2181
foo
bar
Add more brokers: (I failed to get this working yet with the current docker-compose.yml)
docker-compose scale kafka=3
Destroy a cluster:
docker-compose stop
##Note
The default docker-compose.yml
should be seen as a starting point. By default each broker will get a new port number and broker id on restart. Depending on your use case this might not be desirable. If you need to use specific ports and broker ids, modify the docker-compose configuration accordingly, e.g. docker-compose-single-broker.yml:
docker-compose -f docker-compose-single-broker.yml up
##Broker IDs
If you don't specify a broker id in your docker-compose file, it will automatically be generated based on the name that docker-compose gives the container. This allows scaling up and down. In this case it is recommended to use the --no-recreate
option of docker-compose to ensure that containers are not re-created and thus keep their names and ids.
##Automatically create topics
If you want to have kafka-docker automatically create topics in Kafka during
creation, a KAFKA_CREATE_TOPICS
environment variable can be
added in docker-compose.yml
.
Here is an example snippet from docker-compose.yml
:
environment:
KAFKA_CREATE_TOPICS: "Topic1:1:3,Topic2:1:1"
Topic 1
will have 1 partition and 3 replicas, Topic 2
will have 1 partition and 1 replica.
##Tutorial