Prometheus exporter for Druid (http://druid.io/)
Collects HTTP POST JSON data coming from Druid daemons and expose them formatted following the Prometheus standards.
The easiest way to run druid_exporter
is via a virtualenv:
virtualenv .venv
.venv/bin/python setup.py install
.venv/bin/druid_exporter
By default metrics are exposed on TCP port 8000
. Python 2 is not supported.
This exporter is tested and used by the Wikimedia foundation with Druid version 0.11.0, so it might not work as expected for newer versions.
Druid can be configured to emit metrics (JSON data) to a HTTP endpoint configured via the following runtime properties:
http://druid.io/docs/0.11.0/configuration/index.html#http-emitter-module
The druid prometheus exporter accepts HTTP POST data, inspects it and stores/aggregates every supported datapoint into a data structure. It then formats the data on the fly to Prometheus metrics when a GET /metrics is requested.
This exporter is supposed to be run on each host running a Druid daemon.
query/time
[datasource]query/bytes
[datasource]query/cache/total/numEntries
query/cache/total/sizeBytes
query/cache/total/hits
query/cache/total/misses
query/cache/total/evictions
query/cache/total/timeouts
query/cache/total/errors
segment/max
[datasource]segment/count
[datasource]
segment/used
[datasource]segment/scan/pending
segment/assigned/count
[tier]segment/moved/count
[tier]segment/dropped/count
[tier]segment/deleted/count
[tier]segment/unneeded/count
[tier]segment/overShadowed/count
segment/loadQueue/failed
[server]segment/loadQueue/count
[server]segment/dropQueue/count
[server]segment/size
[datasource]segment/unavailable/count
[datasource]segment/underReplicated/count
[datasource, tier]
query/time
[datasource]query/bytes
[datasource]ingest/events/thrownAway
[dataSource]ingest/events/unparseable
[dataSource]ingest/events/processed
[dataSource]ingest/rows/output
[dataSource]ingest/persists/count
[dataSource]ingest/persists/failed
[dataSource]ingest/handoff/failed
[dataSource]ingest/handoff/count
[dataSource]
Realtime metrics have been tested only when emitted by Peons, since the Wikimedia use case (for the moment) is to use Tranquillity rather than Real Time nodes or the Kafka Indexing Service.
Please check the following document for more info: http://druid.io/docs/0.11.0/operations/metrics.html
The JVM metrics are currently not supported, please check other projects like https://github.com/prometheus/jmx_exporter if you need to collect them.
When a Druid cluster is running with multiple coordinators or overlords, only one of them acts as leader and the others are in standby mode. From the metrics emission point of view, this means that only the daemon acting as leader emits druid metrics, the others don't (except for metrics like Jetty ones). This has caused some confusion in our day to day operations, since when a coordinator or a overlord leader looses its status in favor of another one (for example due to a restart) it stops emitting metrics, and the Prometheus exporter keeps reporting the last known state. This might be confusing to see at first (expecially if metrics are aggregated) so the current "fix" is to restart the Druid Prometheus exporter when a coordinator or a overlord leader are restarted. Future versions of this project might contain a fix, pull requests are welcome!