/control-node

🚀 Continuous Delivery and Orchestration as code for Elixir

Primary LanguageElixirMIT LicenseMIT

Control Node

github.com hex.pm hexdocs.pm

🚀 Continuous Delivery and Orchestration as code for Elixir

Installation

def deps do
  [
    {:control_node, "~> 0.5.0"}
  ]
end

Introduction

control_node is an Elixir library which allows developers to build deployment and orchestration workflows as code.

With control_node library you can build your custom orchestration service tailored to your needs i.e. control_node offers APIs to store and manage release tars and deploy releases to remote hosts via SSH and monitor and manage deployed service nodes.

Pre-requisites

In order to use control_node you must ensure the following,

  • Control node should have SSH access all host machines where releases will be deployed
  • Host machines should have EPMD running (it runs by default when an Elixir release is started if you don't change the config)

Features

  • Support multiple namespaces for a release
  • Rollout releases to hosts via SSH
  • Native node monitoring and restart on failover
  • Dynamically scale up/down your release instances
  • Native service monitoring/health check
  • Blue-Green deployment
  • Support failover via heart
  • Support namespace environment variable configuration
  • Rollback releases

Quick example

This library ships with an example service_app under example/ folder. You can try out this library by trying to deploy the release using the following steps,

Clone the repo

$ git clone https://github.com/beamX/control-node
$ cd control-code/

Start an SSH server locally where the release will be deployed,

$ docker-compose up -d

Start iex with distribution turned on

$ iex -S mix
Erlang/OTP 23 [erts-11.0] [source] [64-bit] [smp:8:8] [ds:8:8:10] [async-threads:1] [hipe]

Interactive Elixir (1.10.4) - press Ctrl+C to exit (type h() ENTER for help)
iex(1)> :net_kernel.start([:control_node_test, :shortnames])
iex(control_node_test@hostname)2> 

Execute the Elixir code snippets in the console,

  • Define ServiceApp module (copy paste the code in the console) which will offer API to deploy service_app,
defmodule ServiceApp do
  use ControlNode.Release,
    spec: %ControlNode.Release.Spec{name: :service_app, base_path: "/app/service_app"}
end
  • Declare a host_spec which will hold the details of which host the release can be deployed to
host_spec = %ControlNode.Host.SSH{
  host: "localhost",
  port: 2222,
  user: "linuxserver.io",
  private_key_dir: Path.join([File.cwd!(), "test/fixture", "host-vm/.ssh"])
}
  • Declare a namespace_spec which define the namespace for a given release. Notice that the namespace allows specifying a list of hosts and registry. A registry module offers API to retrieve the release tar and here we use a Local registry which will retrieve the release tar from the filesystem.
namespace_spec = %ControlNode.Namespace.Spec{
  tag: :testing,
  hosts: [host_spec],
  registry_spec: %ControlNode.Registry.Local{path: Path.join(File.cwd!(), "example")},
  deployment_type: :incremental_replace,
  release_cookie: :"YFWZXAOJGTABHNGIT6KVAC2X6TEHA6WCIRDKSLFD6JZWRC4YHMMA===="
}
  • Now we deploy the release to a given namespace_spec i.e. the release we be started on on all the hosts specified in the namespace. Notice that once the deployment is finished control_node_test@hostname automatically connects to release nodes,
{:ok, namespace_manager} = ControlNode.Namespace.start_link(namespace_spec, ServiceApp)
ControlNode.Namespace.deploy(namespace_manager, "0.1.0")
Node.list()

Connect and observe with observer

Once Node.list() shows that the control node is connected to the release nodes then observer can be used to observe and inspect the remote nodes,

l(:observer)
:observer.start()

Real world example

https://github.com/kansi/cnops

Can control node be used to deploy non Elixir/Erlang project?

YES!

https://github.com/kansi/cnops deploys a Golang service hello_go

Under the hood

How it works

  • Upon starting, control_node will try to connect to EMPD process for each specified host and gather info regarding running services on each host
  • In case a service managed by control is already running on a given node control node will retrieve the current running version and start monitoring the release
  • In case no service is found running on a given host, control_node will establish a connection to the host and wait for a deployment command to be issued
  • If any of the monitored service nodes goes down control node will attempt (max. 5) to restart the node

SSH server config to enable tunneling

In order to ensure that Control Node can connect to release node the SSH servers running the release should allow tunneling,

...
AllowTcpForwarding yes
...

Limitations

  • SSH client only supports ed25519 keys. Other keys types are supported only via SSH agent
  • Only short names for nodes are allowed ie. sevice_app@hostname is support and not sevice_app@host1.server.com
  • Nodes of a given release (deployed to different) should have different hostname for eg. if node 1 has node name service_app@host1 then another node of service_app should have a different node name.