ecschedule is a tool to manage ECS Scheduled Tasks.
% ecschedule [dump|apply|run|diff] -conf ecschedule.yaml -rule $ruleName
The ecschedule manages ECS Schedule tasks using a YAML configuration file like following.
region: us-east-1
cluster: clusterName
rules:
- name: taskName1
description: task 1
scheduleExpression: cron(30 15 ? * * *)
taskDefinition: taskDefName
containerOverrides:
- name: containerName
command: [subcommand1, arg]
environment:
HOGE: foo
FUGA: {{ must_env `APP_FUGA` }}
- name: taskName2
description: task2
scheduleExpression: cron(30 16 ? * * *)
taskDefinition: taskDefName2
containerOverrides:
- name: containerName2
command: [subcommand2, arg]
% go install github.com/Songmu/ecschedule/cmd/ecschedule@latest
Action Songmu/ecschedule@main installs ecschedule binary for Linux into /usr/local/bin. This action runs install only.
jobs:
deploy:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v2
- uses: Songmu/ecschedule@main
with:
version: v0.7.1
- run: |
ecschedule -conf ecschedule.yaml apply -all
% ecschedule dump --cluster clusterName --region us-east-1 > ecschedule.yaml
edit and adjust configuration file after it.
% ecschedule -conf ecschedule.yaml apply -rule $ruleName
Before you apply it, you can check the diff in the following way.
% ecschedule -conf ecschedule.yaml diff -rule $ruleName
Execute run
subcommand when want execute arbitrary timing.
% ecschedule -conf ecschedule.yaml run -rule $ruleName
You can use following functions in the configuration file.
env
- expand environment variable or using default value
{{ env "ENV_NAME" "DEFAULT_VALUE" }}
must_env
- expand environment variable
{{ must_env "ENV_NAME" }}
inspired by ecspresso.
tfstate plugin introduces a template function tfstate
.
region: us-east-1
cluster: api
role: ecsEventsRole
rules:
- name: hoge-task-name
description: hoge description
scheduleExpression: cron(0 0 * * ? *)
taskDefinition: task1
group: xxx
platform_version: 1.4.0
launch_type: FARGATE
network_configuration:
aws_vpc_configuration:
subnets:
- {{ tfstate `aws_subnet.private-a.id` }}
- {{ tfstate `aws_subnet.private-c.id` }}
security_groups:
- {{ tfstatef `data.aws_security_group.default['%s'].id` `first` }}
- {{ tfstatef `data.aws_security_group.default['%s'].id` `second` }}
assign_public_ip: ENABLED
containerOverrides:
- name: container1
command: ["subcmd", "argument"]
environment:
HOGE_ENV: {{ env "DUMMY_HOGE_ENV" "HOGEGE" }}
dead_letter_config:
sqs: queue1
propagateTags: TASK_DEFINITION
plugins:
- name: tfstate
config:
path: testdata/terraform.tfstate # path to tfstate file
# or url: s3://my-bucket/terraform.tfstate
{{ tfstate "resource_type.resource_name.attr" }}
will expand to an attribute value of the resource in tfstate.
{{ tfstatef "resource_type.resource_name['%s'].attr" "index" }}
is similar to {{ tfstatef "resource_type.resource_name['index'].attr" }}
.
This function is useful to build a resource address with environment variables.
{{ tfstatef `aws_subnet.ecs['%s'].id` (must_env `SERVICE`) }}
ecschedule is designed to guarantee the uniqueness of job definitions by rule name in the configuration file. This causes the following confusing behavior at the moment. Please be careful.
If the name of a rule that is already reflected in the configuration file is changed, a new rule with that name is created and the old rule with the old name remains, resulting in an unintended double definition. The only solution is to delete the old rule manually.
If a rule is deleted from the configuration file and then reflected, the rule will not be deleted. The only way to do this is to delete this separately, too manually.
This is because ecschedule only manages a subset of cloudwatch event rules, so it cannot distinguish whether a rule name that does not exist in the configuration file is out of ecschedule management or has been deleted.
There should be designs to implement some state management mechanisms to solve these problems. If you have a good solution, I would be happy to make suggestions.