/parseable

Parseable is a log analytics system written in Rust. It is built for high throughput log ingestion cases.

Primary LanguageRustGNU Affero General Public License v3.0AGPL-3.0

Parseable Logo
Cloud native log analytics

Docker Pulls Slack Docs Build

Parseable is a lightweight, cloud native log observability and analytics engine. It is written in Rust and uses Apache Arrow and Parquet.

Parseable uses a simple, index-free mechanism to organize and query data allowing low latency, and high throughput ingestion and query. It can use either a local mount point or object storage (S3/compatible stores) for data storage.

For comparison, Parseable consumes up to ~80% lower memory and ~50% lower CPU than Elastic for similar ingestion throughput. Read more in the benchmarks directory.

🚀 Features

  • Choose your own storage backend - local drive or S3 (or compatible) object store.
  • Ingestion API compatible with HTTP + JSON output of log agents.
  • Query log data with PostgreSQL compatible SQL.
  • Grafana ↗︎ for visualization.
  • Send alerts ↗︎ to webhook targets including Slack.
  • Stats API ↗︎ to track ingestion and compressed data.
  • Single binary includes all components - ingestion, store and query. Built-in UI.

✅ Getting Started

Run the below command to deploy Parseable in local storage mode with Docker.

mkdir -p $HOME/parseable/data
mkdir -p $HOME/parseable/staging

docker run -p 8000:8000 \
  -v $HOME/parseable/data:/parseable/data \
  -v $HOME/parseable/staging:/parseable/staging \
  -e P_FS_DIR=/parseable/data \
  -e P_STAGING_DIR=/parseable/staging \
  parseable/parseable:latest \
  parseable local-store

Once this runs successfully, you'll see dashboard at http://localhost:8000. You can login to the dashboard default credentials admin, admin.

Send log events

curl --location --request POST 'http://localhost:8000/api/v1/ingest' \
--header 'X-P-META-meta1: value1' \
--header 'X-P-TAG-tag1: value1' \
--header 'X-P-Stream: demo' \
--header 'Authorization: Basic YWRtaW46YWRtaW4=' \
--header 'Content-Type: application/json' \
--data-raw '[
    {
        "id": "434a5f5e-2f5f-11ed-a261-0242ac120002",
        "datetime": "24/Jun/2022:14:12:15 +0000",
        "host": "153.10.110.81",
        "user-identifier": "Mozilla/5.0 Gecko/20100101 Firefox/64.0",
        "method": "PUT",
        "status": 500,
        "referrer": "http://www.google.com/"
    }
]'

Note: The X-P-Stream header is used to specify the log stream where data is sent. Parseable will create the stream automatically if it doesn't exist already.

Query the stream

You can see the events in Parseable UI, or use the below curl command to see the query response on CLI.

NOTE: Please change the startTime and endTime to the time range corresponding to the event you sent in the previous step.

curl --location --request POST 'http://localhost:8000/api/v1/query' \
--header 'Authorization: Basic YWRtaW46YWRtaW4=' \
--header 'Content-Type: application/json' \
--data-raw '{
    "query":"select * from demo",
    "startTime":"2023-06-09T00:00:00+00:00",
    "endTime":"2023-06-09T23:59:00+00:00"
}'

📈 Benchmarking

Parseable is benchmarked with K6. Please find the results and details on how to run the benchmark in your environment in the benchmarks directory.

📚 Documentation

🎯 Motivation

Traditionally, logging has been seen as a text search problem. Log volumes were not high, and data ingestion or storage were not really issues. This led us to today, where all the logging platforms are primarily text search engines.

But with log data growing exponentially, today's log data challenges involve whole lot more – Data ingestion, storage, and observation, all at scale. We are building Parseable to address these challenges.

🩺 Support

🏆 Contributing

Refer to the contributing guide here ↗︎.

Contributors

Supported by