PostgreSQL | MySQL version control engine. Applies effects deliberately.
Migrate your PostgreSQL or MySQL database from Erlang code with no effort. This amazing toolkit has one and only purpose - consistently upgrade database schema, using Erlang stack and plain SQL. Feel free to run it with any PostgreSQL/MySQL Erlang library (and see several ready-to-use examples below). As an extra - do this in "no side-effects" mode.
- Current limitations
- Quick start
- "No-effects" approach and tools used to achieve it
- Functional programming abstractions used
- License
- up transactional migration available only. No downgrade calls available. Either whole up migration completes OK or failed and rolled back to the state before migration.
- Validated MySQL implementation obviously featured with implicit commit behavior, which means that truly transactional MySQL upgrades limited in scope. At the same time you may adjust MySQL transaction callback, as it is proposed by API.
- migrations engine deliberately isolated from any specific database library. This way engine user is free to choose from variety of frameworks (see tested combinations here) and so on.
Just call pure_migrations:migrate/3
(see specification here), providing:
Path
to migration scripts folder (strictly and incrementally enumerated).FTx
transaction handlerFQuery
database queries execution handler
Migration logic is idempotent and could be executed multiple times against the same database with the same migration scripts set. Moreover, it is safe to migrate your database concurrently (as a part of nodes startup in scalable environments and if you providing proper transaction handler). Please see verified integrations and live code snippets below.
All integrations validated against PostgreSQL 9.4/9.6
Database dialect | Library | Example |
---|---|---|
postgres | epgsql/epgsql:4.2.1 | epgsql test |
postgres | semiocast/pgsql:v26.0.2 | spgsql test |
postgres | processone/p1_pgsql:1.1.6 | p1pgsql test |
mysql | mysql-otp/mysql-otp:1.4.0 | otp_mysql test |
postgres | any library with basic sql functional | generic test |
Sure! Please follow these simple steps below:
- for local build just amend related PostgreSQL
or MySQL images references inside project Makefile
and just run
make local
. - for your CI build experiments please follow related Travis docs for Postgres or MySQL instructions.
Library production code has no third-party dependencies at all. Code becomes extremely lightweight and decoupled from particular library bottlenecks. User absolutely free to choose any implementation (maybe one of validated ones) and it's version as well.
The approach could be expressed as 2 rules:
- Each script name has a number prefix
- Numbers should start from 0 and increment strictly by 1
This model gives much more clarity for migrations sequence. And there is no chance to interlace migrations accidentally. Which is the case if multiple migrations are being developed and merged to default branch simultaneously.
PostgreSQL and epgsql/epgsql
- most popular out of onboarded postgres integrations
- transactions with proper stack trace available out of the box
- reasonably structured query responses, provided with data and its schema
- although binary strings could be very reasonable, sometimes code too verbose because of this
Click to expand
migrate() ->
Conf = ?config(super_account),
Path = ?env(scripts_path),
{ok, Conn} = epgsql:connect(Conf),
MigrationCall =
pure_migrations:migrate(
Path,
fun(F) -> epgsql:with_transaction(Conn, fun(_) -> F() end) end,
fun(Q) ->
case epgsql:squery(Conn, Q) of
{ok, [{column, <<"version">>, _, _, _, _, _, _, _},
{column, <<"filename">>, _, _, _, _, _, _, _}], []} ->
[];
{ok, [{column, <<"version">>, _, _, _, _, _, _, _},
{column, <<"filename">>, _, _, _, _, _, _, _}], Data} ->
[{list_to_integer(binary_to_list(BinV)), binary_to_list(BinF)} || {BinV, BinF} <- Data];
{ok, [{column, <<"max">>, _, _, _, _, _, _, _}], [{null}]} ->
% It has to be -1 or it will get an error during initialization
-1;
{ok, [{column, <<"max">>, _, _, _, _, _, _, _}], [{N}]} ->
% The version number is stored in the int4 type and ranges from -2,147,483,648 to 2,147,483,647
list_to_integer(binary_to_list(N));
{ok, [
{column, <<"version">>, _, _, _, _, _},
{column, <<"filename">>, _, _, _, _, _}], Data} ->
[{list_to_integer(binary_to_list(BinV)), binary_to_list(BinF)} || {BinV, BinF} <- Data];
{ok, [{column, <<"max">>, _, _, _, _, _}], [{null}]} -> -1;
{ok, [{column, <<"max">>, _, _, _, _, _}], [{N}]} ->
list_to_integer(binary_to_list(N));
{ok, _, _} -> ok;
{ok, _} -> ok;
Default ->
% Match multiple SQL statements in a script
Res = priv_is_valid(Default),
case Res of
true->
ok;
_ ->
Default
end
end
end),
% ...
%% more preparation steps if needed
% ...
%% migration call
Res = MigrationCall(),
% imboy_log:debug(io:format("~p~n", [Res])),
ok = epgsql:close(Conn),
Res.
priv_is_valid(List) ->
lists:all(fun(E) ->
case E of
{ok, _} -> true;
{ok, _, _} -> true;
_ -> false
end
end, List).
Also see examples from live epgsql integration tests here
PostgreSQL and semiocast/pgsql
- no need for extra parsing (strings, numbers, ...)
- queries results structure has no metadata, like column types or names, which could be sub-optimal sometimes
- no transactions out of the box
Click to expand
Conn = ?config(conn, Opts),
MigrationCall =
pure_migrations:migrate(
"scripts/folder/path",
fun(F) ->
pgsql_connection:simple_query("BEGIN", Conn),
try F() of
Res ->
pgsql_connection:simple_query("COMMIT", Conn),
Res
catch
_:Problem ->
pgsql_connection:simple_query("ROLLBACK", Conn),
{rollback, Problem}
end
end,
fun(Q) ->
case pgsql_connection:simple_query(Q, Conn) of
{{select, 0}, []} -> [];
{{select, 1}, Data = [{_V, _F}|_]} ->
[{V, binary_to_list(BinF)} || {V, BinF} <- Data];
{{select, 1}, [{null}]} -> -1;
{{select, 1}, [{N}]} -> N;
{{insert, 0, 1}, []} -> ok;
{{create, table},[]} -> ok;
{error, Details} -> {error, Details};
_ -> ok
end
end),
...
%% more preparation steps if needed
...
%% migration call
ok = MigrationCall(),
Also see examples from live semiocast/pgsql integration tests here
PostgreSQL and processone/p1_pgsql
- least popular lib,but at the same time - most succinct in terms of integration code (see below)
- decent types balance gives opportunity to keep code clean
- no transactions out of the box
- error reporting different for postgres 9.4/9.6
Click to expand
Conn = ?config(conn, Opts),
MigrationCall =
pure_migrations:migrate(
"scripts/folder/path",
fun(F) ->
pgsql:squery(Conn, "BEGIN"),
try F() of
Res ->
pgsql:squery(Conn, "COMMIT"),
Res
catch
_:Problem ->
pgsql:squery(Conn, "ROLLBACK"),
{rollback, Problem}
end
end,
fun(Q) ->
case pgsql:squery(Conn, Q) of
{ok, [{error, Details}]} -> {error, Details};
{ok, [{_, [
{"version", text, _, _, _, _, _},
{"filename", text, _, _, _, _, _}], Data}]} ->
[{list_to_integer(V), F} || [V, F] <- Data];
{ok, [{"SELECT 1", [{"max", text, _, _, _, _, _}], [[null]]}]} -> -1;
{ok, [{"SELECT 1", [{"max", text, _, _, _, _, _}], [[N]]}]} ->
list_to_integer(N);
{ok, _} -> ok
end
end),
...
%% more preparation steps if needed
...
%% migration call
ok = MigrationCall(),
Also see examples from live epgsql integration tests here
MySQL and mysql-otp/mysql-otp
- almost no result-set parsing required
- implicit commit specifics a kind an obstacle for simple and safe migration
- mysql docker tooling should be operated carefully and ensured for proper startup before any use
Click to expand
Conn = ?config(conn, Opts),
MigrationCall =
pure_migrations:migrate(
"scripts/folder/path",
fun(F) ->
%% no full-scope tx API available here
%% alternatively use mysql:transaction/2, but please be aware about
%% mysql implicit transactions commit behavior
try F() of
Res -> Res
catch
_:Problem -> {rollback_unavailable, Problem}
end
end,
fun(Q) ->
case mysql:query(Conn, Q) of
{error, Details} -> {error, Details};
{ok,[<<"version">>,<<"filename">>],[]} -> [];
{ok,[<<"version">>,<<"filename">>], Data} ->
[{V, binary_to_list(F)} || [V, F] <- Data];
{ok,[<<"max(version)">>],[[null]]} -> -1;
{ok,[<<"max(version)">>],[[V]]} -> V;
{ok, _} -> ok;
ok -> ok
end
end),
...
%% more preparation steps if needed
...
%% migration call
ok = MigrationCall(),
Also see examples from live epgsql integration tests here
Oh, there is more! Library implemented in the way, that all side-effects either externalized or deferred explicitly. Reasons are quite common:
- bring side-effects as close to program edges as possible. Which may mean enhanced code reasoning, better bugs reproduceability, etc...
- simplify module contracts testing
- library users empowered to re-run idempotent code safely. Well, if tx/query handlers are real ones - execution is still idempotent (at application level) and formally pure. But purity maintained inside library code only. One call is to be issued anyway - migrations table creation, if this one does not exists.
There are 2 externalized kind of effects:
- transaction management handler
- database queries handler Although, those two can`t be pure in real application, it is fairly simple to replace them with their pure versions if we would like to (for debug purposes, or testing, or something else).
Other effects (like file operations) are deferred in bulk with outcome like:
- pure referentially-transparent program actions composed only. Impact or any communication with external world postponed until later stages
- library users decide when they ready to apply migration changes. Maybe for some reason they would like to prepare execution -> prepare migrations folder content -> run migrations.
This trick is quite useful if someone would like to compose two functions without their actual execution (or without their application, alternatively speaking). This pretty standard routine may look like below (Scala or Kotlin+Arrow):
val divideByTwo = (number : Int) => number / 2;
val addThree = (number: Int) => number + 3;
val composed = addThree compose divideByTwo
Simplistic Erlang version:
compose(F1, F2) -> fun() -> F2(F1()) end.
There area few places in library with clear need to compose function A and another function B inside deferred execution context. Specifics is that A supplies list of objects, and B should be applied to each of them. Sounds like some functor B to be applied to A output, when this output is being wrapped into future execution context. Two cases of this appeared in library:
- have functor running and produce nested list of contexts:
%% Map/1 call here produces new context (defferred function call)
map(Generate, Map) -> fun() -> [Map(R) || R <- Generate()] end.
- flatten (or fold) contexts (or function calls) list to a single one:
flatten(Generate) -> fun() -> [ok = R() || R <- Generate()], ok end.
Partial application is very useful, in case if not all function arguments known yet. Or maybe there is deliberate decision to pass some of arguments later on. Again, in Scala it may look like:
val add = (a: Int, b: Int) => a + b
val partiallyApplied = add(3, _)
Library code has very simplistic partial application, done for exact arguments number (although it is easy to generalize it for arguments, represented as list):
Partial = fun(V_F) -> do_migration(Path, FQuery, V_F) end,
MIT, see LICENSE for more details.