Event store implemented in Elixir. Uses PostgreSQL as the underlying storage engine.
Requires Elixir v1.6 and PostgreSQL v9.5 or newer.
EventStore supports running on a cluster of nodes.
MIT License
This README and the following guides follow the
master
branch which may not be the currently published version. Read docs for the latest published version of EventStore on Hex.
- Getting started
- Using the EventStore
- Subscriptions
- Running on a cluster
- Event serialization
- Upgrading an EventStore
- Used in production?
- Backup and administration
- Benchmarking performance
- Contributing
- Need help?
Define an event store module:
defmodule MyEventStore do
use EventStore, otp_app: :my_app
end
Start the event store:
{:ok, _pid} = MyEventStore.start_link()
Create one or more event structs to be persisted (serialized to JSON by default):
defmodule ExampleEvent do
defstruct [:key]
end
Append events to a stream:
stream_uuid = UUID.uuid4()
expected_version = 0
events = [
%EventStore.EventData{
event_type: "Elixir.ExampleEvent",
data: %ExampleEvent{key: "value"},
metadata: %{user: "someuser@example.com"}
}
]
:ok = MyEventStore.append_to_stream(stream_uuid, expected_version, events)
Read all events from a single stream, starting at the stream's first event:
{:ok, events} = MyEventStore.read_stream_forward(stream_uuid)
More: Usage guide
Subscribe to events appended to all streams:
{:ok, subscription} = MyEventStore.subscribe_to_all_streams("example_subscription", self())
# Wait for the subscription confirmation
receive do
{:subscribed, ^subscription} ->
IO.puts("Successfully subscribed to all streams")
end
# Receive a batch of events appended to the event store
receive do
{:events, events} ->
IO.puts("Received events: #{inspect events}")
# Acknowledge successful receipt of events
:ok = MyEventStore.ack(subscription, events)
end
In production use you would use a GenServer
subscriber process and the handle_info/2
callback to receive events.
More: Subscriptions guide
Yes, this event store is being used in production.
PostgreSQL is used for the underlying storage. Providing guarantees to store data securely. It is ACID-compliant and transactional. PostgreSQL has a proven architecture. A strong reputation for reliability, data integrity, and correctness.
You can use any standard PostgreSQL tool to manage the event store data:
Run the benchmark suite using mix with the bench
environment, as configured in config/bench.exs
. Logging is disabled for benchmarking.
MIX_ENV=bench mix do es.reset, app.start, bench
Example output:
## AppendEventsBench
benchmark name iterations average time
append events, single writer 100 20288.68 µs/op
append events, 10 concurrent writers 10 127416.90 µs/op
append events, 20 concurrent writers 5 376836.60 µs/op
append events, 50 concurrent writers 2 582350.50 µs/op
## ReadEventsBench
benchmark name iterations average time
read events, single reader 500 3674.93 µs/op
read events, 10 concurrent readers 50 44653.98 µs/op
read events, 20 concurrent readers 20 73927.55 µs/op
read events, 50 concurrent readers 10 188244.80 µs/op
## SubscribeToStreamBench
benchmark name iterations average time
subscribe to stream, 1 subscription 100 27687.97 µs/op
subscribe to stream, 10 subscriptions 50 56047.72 µs/op
subscribe to stream, 20 subscriptions 10 194164.40 µs/op
subscribe to stream, 50 subscriptions 5 320435.40 µs/op
After running two benchmarks you can compare the runs:
MIX_ENV=bench mix bench.cmp -d percent
You can also produce an HTML page containing a graph comparing benchmark runs:
MIX_ENV=bench mix bench.graph
Taking the above example output, the append events benchmark is for writing 100 events in a single batch. That's what the µs/op average time is measuring. For a single writer it takes on average 0.02s per 100 events appended (4,929 events/sec) and for 50 concurrent writers it's 50 x 100 events in 0.58s (8,586 events/sec).
For reading events it takes a single reader 3.67ms to read 100 events (27,211 events/sec) and for 50 concurrent readers it takes 0.19s (26,561 events/sec).
The purpose of the benchmark suite is to measure the performance impact of proposed changes, as opposed to looking at the raw numbers. The above figures are taken when run against a local PostgreSQL database. You can run the benchmarks against your own hardware to get indicative performance figures for the Event Store.
The benchmark suite is configured to use Erlang's external term format serialization. Using another serialization format, such as JSON, will likely have a negative impact on performance.
Tests can be run using any Postgres database instance, including via Docker.
To use Docker, first pull the latest Postgres image:
docker pull postgres
A tmpfs mount can be used to run the Docker container with the Postgres data directory stored in memory.
docker run --rm \
--name postgres \
--tmpfs=/pgtmpfs \
-e PGDATA=/pgtmpfs \
-e POSTGRES_PASSWORD=postgres \
-e POSTGRES_USER=postgres \
-p 5432:5432 \
postgres
Create and initialize the test event store databases:
MIX_ENV=test mix event_store.setup
Run the test suite:
mix test
Pull requests to contribute new or improved features, and extend documentation are most welcome.
Please follow the existing coding conventions, or refer to the Elixir style guide.
You should include unit tests to cover any changes.
EventStore exists thanks to the following people who have contributed.
- Andreas Riemer
- Andrey Akulov
- Basile Nouvellet
- Ben Smith
- Bruce Williams
- Chris Brodt
- Chris Martin
- Christian Green
- Craig Savolainen
- David Soff
- Derek Kraan
- Diogo Scudelletti
- Dominik Guzei
- Douglas Vought
- Eamon Taaffe
- Floris Huetink
- Jan Vereecken
- Kai Kuchenbecker
- Kaz Walker
- Morten Berg Nissen
- Nicholas Henry
- Olafur Arason
- Ole Michaelis
- Paul Iannazzo
- Raphaël Lustin
- Samuel Roze
- Simon Harris
- Stuart Corbishley
- Thomas Coopman
- Victor Oliveira Nascimento
- Yamil Díaz Aguirre
- Yannis Weishaupt
Please open an issue if you encounter a problem, or need assistance.
For commercial support, and consultancy, please contact Ben Smith.