/hivemq-docker-images

This repository contains the Dockerfiles for the HiveMQ Base image and HiveMQ DNS Discovery image, as found at https://hub.docker.com/r/hivemq/hivemq3/

Primary LanguageDockerfile

Table of Contents

HiveMQ Docker images

This repository provides the Dockerfile and context for the images hosted in the HiveMQ Docker Hub repository.

DNS Discovery Image

The HiveMQ DNS discovery image comes with a DNS discovery plugin. It can be used with any container orchestration engine that supports service discovery using a round-robin A record.

A custom solution supplying the A record could be used as well.

The following environment variables should be used to customize the discovery and broker configuration respectively.

Environment Variable Default value Meaning
HIVEMQ_DNS_DISCOVERY_ADDRESS - Address to get the A record that will be used for cluster discovery
HIVEMQ_DNS_DISCOVERY_INTERVAL 31 Discovery interval in seconds
HIVEMQ_DNS_DISCOVERY_TIMEOUT 30 DNS resolution wait time in seconds
HIVEMQ_CLUSTER_PORT 8000 Port used for cluster transport
HIVEMQ_LICENSE - base64 encoded license file to use for the broker
HIVEMQ_BIND_ADDRESS - Set the cluster transport bind address, only necessary if the default policy (resolve hostname) fails
HIVEMQ_WEB_UI_USER admin Set the username for the Web UI login
HIVEMQ_WEB_UI_PASSWORD SHA256 of adminhivemq (default) Set the password hash for Web UI authentication

Building

To build the image, you must first obtain the HiveMQ DNS discovery plugin and copy the jar file to the hivemq3/dns-image folder.

The image can then be built by running docker build -t hivemq-dns . in the hivemq3/dns-image folder.

Usage

Following are two examples, describing how to use this image on Docker Swarm and Kubernetes respectively.

Other environments (provided they support DNS discovery in some way) are compatible as well.

Docker Swarm

Please note that using Docker Swarm in production is not recommended.

  • Start a single node Swarm cluster by running:
docker swarm init
  • Create an overlay network for the cluster nodes to communicate on:
docker network create -d overlay --attachable myNetwork
  • Create the HiveMQ service on the network
docker service create \
  --replicas 3 --network myNetwork \
  --env HIVEMQ_DNS_DISCOVERY_ADDRESS=tasks.hivemq \
  --publish target=1883,published=1883 \
  --publish target=8080,published=8080 \
  -p 8000:8000/udp \
  --name hivemq \
    hivemq/hivemq3:dns-latest

This will provide a 3 node cluster with the MQTT(1883) and Web UI(8080) ports forwarded to the host network.

This means you can connect MQTT clients on port 1883. The connection will be forwarded to any of the cluster nodes.

The HiveMQ Web UI can be used in a single node cluster. A sticky session for the HTTP requests in clusters with multiple nodes cannot be upheld with this configuration, as the internal load balancer forwards requests in an alternating fashion. To use sticky sessions the Docker Swarm Enterprise version is required.

Managing the cluster

To scale the cluster up to 5 nodes, run

docker service scale hivemq=5

To remove the cluster, run

docker service rm hivemq

To read the logs for all HiveMQ nodes in real time, use

docker service logs hivemq -f

To get the log for a single node, get the list of service containers using

docker service ps hivemq

And print the log using

docker service logs <id>

where <id> is the container ID listed in the service ps command.

Kubernetes

On Kubernetes, an appropriate deployment configuration is necessary to utilize DNS discovery. A headless service will provide a DNS record for the broker that can be used for discovery.

Following is an example configuration for a HiveMQ cluster with 3 nodes using DNS discovery in a replication controller setup.

Please note that you may have to replace HIVEMQ_DNS_DISCOVERY_ADDRESS according to your Kubernetes namespace and configured domain.

apiVersion: v1
kind: ReplicationController
metadata:
  name: hivemq-replica
spec:
  replicas: 3
  selector:
    app: hivemq-cluster1
  template:
    metadata:
      name: hivemq-cluster1
      labels:
        app: hivemq-cluster1
    spec:
      containers:
      - name: hivemq-pods
        image: hivemq/hivemq3:dns-latest
        ports:
        - containerPort: 8080
          protocol: TCP
          name: web-ui
        - containerPort: 1883
          protocol: TCP
          name: mqtt
        env:
        - name: HIVEMQ_DNS_DISCOVERY_ADDRESS
          value: "hivemq-discovery.default.svc.cluster.local."
        - name: HIVEMQ_DNS_DISCOVERY_TIMEOUT
          value: "20"
        - name: HIVEMQ_DNS_DISCOVERY_INTERVAL
          value: "21"
        readinessProbe:
          tcpSocket:
            port: 1883
          initialDelaySeconds: 30
          periodSeconds: 60
          failureThreshold: 60
        livenessProbe:
          tcpSocket:
            port: 1883
          initialDelaySeconds: 30
          periodSeconds: 60
          failureThreshold: 60
---
kind: Service
apiVersion: v1
metadata:
  name: hivemq-discovery
  annotations:
    service.alpha.kubernetes.io/tolerate-unready-endpoints: "true"
spec:
  selector:
    app: hivemq-cluster1
  ports:
    - protocol: TCP
      port: 1883
      targetPort: 1883
  clusterIP: None

Accessing the Web UI

To access the HiveMQ Web UI for a cluster running on Kubernetes, follow these steps:

  • Create a service exposing the Web UI of the HiveMQ service. Use the following YAML definition:
kind: Service
apiVersion: v1
metadata:
  name: hivemq-web-ui
spec:
  selector:
    app: hivemq-cluster1
  ports:
    - protocol: TCP
      port: 8080
      targetPort: 8080
  sessionAffinity: ClientIP
  type: LoadBalancer
  • Create the service using kubectl create -f web.yaml

Note that depending on your provider of Kubernetes environment, load balancers might not be available or additional configuration may be necessary to access the Web UI.

Accessing the MQTT port using external clients

To allow access for the MQTT port of a cluster running on Kubernetes, follow these steps:

  • Create a service exposing the MQTT port using a load balancer. You can use the following YAML definition:
kind: Service
apiVersion: v1
metadata:
  name: hivemq-mqtt
  annotations:
    service.spec.externalTrafficPolicy: Local
spec:
  selector:
    app: hivemq-cluster1
  ports:
    - protocol: TCP
      port: 1883
      targetPort: 1883
  type: LoadBalancer

Note that the externalTrafficPolicy annotation is necessary to allow the Kubernetes service to maintain a larger amount of concurrent connections.
See Source IP for Services for more information.

Setting the Web UI username and password

The environment variable HIVEMQ_WEB_UI_PASSWORD allows you to set the password of the Web UI by defining a SHA256 hash for a custom password. Additionally, you can also configure the username, using the environment variable HIVEMQ_WEB_UI_USER

See Generate a SHA256 Password to read more about how to generate the password hash.

Adding a license

To use a license with this image, you must first encode it as a string.

To do so, run cat license.lic | base64 (replace license.lic with the path to your license file).

Set the resulting string as the value for the HIVEMQ_LICENSE environment variable of the container.

Overriding the bind address

By default this image will attempt to set the bind address using the containers ${HOSTNAME} to ensure that HiveMQ will bind the cluster connection to the correct interface so a cluster can be formed.

This behavior can be overridden by setting any value for the environment variable HIVEMQ_BIND_ADDRESS. The broker will attempt to use the given value as the bind address instead.

HiveMQ base image

The HiveMQ base image installs and optimizes the HiveMQ installation for execution as a container.

It is meant to be used to build custom images or to run a dockerized HiveMQ locally for testing purposes.

Tags

The repository on the HiveMQ Docker Hub repository provides different versions of the HiveMQ image using tags:

Tag Meaning
latest This tag will always point to the latest version of the base image
dns-latest This tag will always point to the latest version of the DNS discovery image
<version> Base image providing the given version of the broker (e.g. 3.4.1)
dns-<version> DNS discovery image based on the given version base image