/blambda

Blambda! is a custom runtime for AWS Lambda that lets you write functions using Babashka

Primary LanguageClojureOtherNOASSERTION

Blambda

Blambda is a custom runtime for AWS Lambda that lets you write functions using Babashka. It is based on the fantastic work that Tatu Tarvainen did on taking care of the heavy lifting of interacting with the Lambda runtime API to process function invocations in bb-lambda. I'm using the bootstrap.clj from that project directly, and have just rewritten the machinery around it to remove Docker in favour of zip files, which I think are simpler (but maybe not easier).

Blambda also owes a huge debt to Karol Wójcik's awesome Holy Lambda, which is a full-featured and production-grade runtime for Clojure on AWS Lambda. I've read a lot of Holy Lambda code to figure out how to do the complicated bits of Babashka-ing on lambda. 💜

Using Blambda

Blambda is meant to be used as a library from your Babashka project. The easiest way to use it is to add tasks to your project's bb.edn.

This example assumes a basic bb.edn like this:

{:deps {net.jmglov/blambda
        {:git/url "https://github.com/jmglov/blambda.git"
         :git/tag "v0.2.0"
         :git/sha "fa394bb"}}
 :tasks
 {:requires ([blambda.cli :as blambda])
  blambda {:doc "Controls Blambda runtime and layers"
           :task (blambda/dispatch
                  {:deps-layer-name "hello-deps"
                   :lambda-name "hello"
                   :lambda-handler "hello/hello"
                   :lambda-iam-role "arn:aws:iam::123456789100:role/hello-lambda"
                   :source-files ["hello.clj"]})}}}

and a simple lambda function contained in a file src/hello.clj looking like this:

(ns hello)

(defn hello [{:keys [name] :or {name "Blambda"} :as event} context]
  (prn {:msg "Invoked with event",
        :data {:event event}})
  {:greeting (str "Hello " name "!")})

You can find more examples in the examples directory in this repo.

Building

Custom runtime layer

To build Blambda with the default Babashka version and platform, run:

bb blambda build-runtime-layer

To see what the default Babashka version and platform are, run:

bb blambda build-runtime-layer --help

To build a custom runtime with Babashka 1.1.173 on amd64, run:

bb blambda build-runtime-layer --bb-version 1.1.173 --bb-arch arm64

Dependencies

All but the most basic lambda functions will depend on Clojure libraries. Blambda has support for keeping these dependencies in a separate layer so that your function deployment contains only the source of your lambda itself.

Your lambda should declare its dependencies in bb.edn or deps.edn as normal; for example, a lambda function that interacts with S3 using awyeah-api might have a src/bb.edn that looks like this:

{:paths ["."]
 :deps {com.cognitect.aws/endpoints {:mvn/version "1.1.12.206"}
        com.cognitect.aws/s3 {:mvn/version "822.2.1109.0"}
        com.grzm/awyeah-api {:git/url "https://github.com/grzm/awyeah-api"
                             :git/sha "0fa7dd51f801dba615e317651efda8c597465af6"}
        org.babashka/spec.alpha {:git/url "https://github.com/babashka/spec.alpha"
                                 :git/sha "433b0778e2c32f4bb5d0b48e5a33520bee28b906"}}}

To build your dependencies layer:

bb blambda build-deps-layer

Lambda

To build your lambda artifact:

bb blambda build-lambda

Deploying

To deploy Blambda using Terraform (recommended), first write the config files:

bb blambda terraform write-config

If you'd like to use S3 to store your lambda artifacts (layers and lambda zipfile), run:

bb blambda terraform write-config \
  --use-s3 --s3-bucket BUCKET --s3-artifact-path PATH

Replace BUCKET and PATH with the appropriate bucket and path. If the bucket you specify doesn't exist, it will be created by Terraform the first time you deploy. If you want to use an existing bucket, you'll need to import the bucket after you generate your Terraform config:

bb blambda terraform import-artifacts-bucket --s3-bucket BUCKET

You can also include extra Terraform configuration. For example, if you want to create an IAM role for your lambda, you might have a file called tf/iam.tf in your repo which looks something like this:

resource "aws_iam_role" "hello" {
  name = "site-analyser-lambda"

  assume_role_policy = jsonencode({
    Version = "2012-10-17"
    Statement = [
      {
        Action = "sts:AssumeRole"
        Effect = "Allow"
        Principal = {
          Service = "lambda.amazonaws.com"
        }
      }
    ]
  })
}

resource "aws_iam_policy" "hello" {
  name = "site-analyser-lambda"

  policy = jsonencode({
    Version = "2012-10-17"
    Statement = [
      {
        Effect = "Allow"
        Action = [
          "s3:GetObject"
        ]
        Resource = [
          "arn:aws:s3:::logs.jmglov.net/logs/*"
        ]
      },
      {
        Effect = "Allow"
        Action = [
          "logs:CreateLogStream",
          "logs:PutLogEvents"
        ]
        Resource = "${aws_cloudwatch_log_group.lambda.arn}:*"
      },
      {
        Effect = "Allow",
        Action = [
          "s3:ListBucket"
        ]
        Resource = [
          "arn:aws:s3:::logs.jmglov.net"
        ]
        Condition = {
          "StringLike": {
            "s3:prefix": "logs/*"
          }
        }
      }
    ]
  })
}

resource "aws_iam_role_policy_attachment" "hello" {
  role = aws_iam_role.hello.name
  policy_arn = aws_iam_policy.hello.arn
}

Note how you can refer to resources defined by Blambda, for example ${aws_cloudwatch_log_group.lambda.arn}. You can see what resources are defined by looking at resources/blambda.tf and resources/lambda_layer.tf.

You can now use this IAM role with your lambda:

bb blambda terraform write-config \
  --extra-tf-config tf/iam.tf \
  --lambda-iam-role '${aws_iam_role.hello.arn}'

To deploy, run:

bb blambda terraform apply

To deploy an arm64 runtime so that you can use AWS Graviton 2 lamdbas (which AWS say will give you up to "34%" better price performance), run:

bb blambda build-runtime-layer --bb-arch arm64 && \
bb blambda terraform write-config --bb-arch arm64 && \
bb blambda terraform apply

Note that if you do this, you must configure your lambda as follows:

  • Runtime: Custom runtime on Amazon Linux 2
  • Architecture: arm64

If you prefer not to use Terraform, you can use the AWS CLI, CloudFormation, or the AWS console, but Blambda doesn't currently offer any tooling for those options.

Using pods

If you want to use one of the pods in Babashka's pods registry, you should add it to your dependencies file as normal, using the :pods key. For example, to use the tzzh/aws pod, your src/bb.edn should look like this:

{:paths ["."]
 :pods {tzzh/aws {:version "0.0.3"}}}

Since your lambda won't be invoked as a project, you won't be able to use the nice automatic pod loading as described in the Babashka pods documentation, so you'll need to load the exact pod version you declared in your dependencies. For example:

(ns s3
  (:require [babashka.pods :as pods]))

(pods/load-pod 'tzzh/aws "0.0.3")
(require '[pod.tzzh.s3 :as s3])

This is gross and I'm sorry. 😢

There is a full example of a lambda using a pod in examples/s3-lister-pod.

Using a local pod

Babashka also supports using pods from the local filesystem. To do this in Blambda, make sure your pod executable exists in your src directory, then declare it in your src/bb.edn like so:

{:paths ["."]
 :pods {tzzh/aws {:path "./my-pod"}}}

Blambda will copy this pod into your deps layer at the root, meaning it will end up in /opt in the deployed lambda image. The Blambda runtime adds /opt to the PATH environment variable when invoking bb, and Babashka will look up pods on the local filesystem using PATH, so you can load the pod in your lambda simply by referring to it by its name:

(pods/load-pod "my-pod")