An example of how to deploy Apache Airflow on Amazon ECS Fargate.
The purpose of this project is to demonstrate how to deploy Apache Airflow on AWS Elastic Container Service using the Fargate capacity provider. The code in this repository is meant as an example to assist programmers create their own configuration. However, one can deploy it using the steps described in Setup an ECS cluster.
Airflow and ECS have many features and configuration options. This project covers many use cases. For example:
- autoscale workers to zero
- route airflow service logs to CloudWatch and to Kinesis Firehose using fluentbit
- use remote_logging to send/receive worker logs to/from S3
- use the AWS provider SecretsManagerBackend to store/consume sensitive configuration options in SecretsManager
- run a single command as standalone ECS task (eg.
airflow db init
) - get a shell into a running container via ECS exec
- send Airflow statsd metrics to CloudWatch
These configuration examples should prove helpful even to those who aren't running Airflow on ECS.
Please see the following tree for a description of the main directories/files. This layout is not based on any standard. One could move the contents of scripts
into deploy_airflow_on_ecs_fargate
. Files named *_config.py
could be placed in a separate config
directory. The location of a file is less important than the quality of the code inside it.
├── build .............................. anything related to building container images
│ ├── dev ............................ development config referenced by docker-compose.yml
│ └── prod ........................... production config used to build image sent to ECR
├── dags ............................... AIRFLOW_HOME/dags directory
├── deploy_airflow_on_ecs_fargate ...... arbitrary python package import-able from dags/plugins used to store config files / extra python modules
│ ├── celery_config.py ............... custom celery configuration
│ └── logging_config.py .............. custom logging configuration
├── docker-compose.yml ................. development environment build config
├── infrastructure ..................... ECS terraform configuration
│ ├── terraform.tfvars.template ...... a template variables file for defining sensitive information required to deploy infrastructure
│ └── *.tf ........................... example ECS cluster terraform configuration
├── plugins ............................ AIRFLOW_HOME/plugins directory
└── scripts
├── put_airflow_worker_xxx.py ...... script used by airflow_metrics ECS service to send custom autoscaling metrics to cloudwatch
└── run_task.py .................... an example python script for running standalone tasks on the ECS cluster
- Initialize the metadata db
docker compose run --rm airflow-cli db init
- Create an admin user
docker compose run --rm airflow-cli users create --email airflow@example.com --firstname airflow --lastname airflow --password airflow --username airflow --role Admin
- Start all services
docker compose up -d
- Initialize the terraform directory
terraform -chdir=infrastructure init
- (Optional) Create a
terraform.tfvars
file and set the variablesaws_region
,metadata_db
andfernet_key
cp infrastructure/terraform.tfvars.template infrastructure/terraform.tfvars
- Create the ECR repository to store the custom airflow image.
terraform -chdir=infrastructure apply -target=aws_ecr_repository.airflow
- Obtain the repository URI via
awscli
or the AWS console.
aws ecr describe-repositories
{
"repositories": [
{
"repositoryArn": "arn:aws:ecr:us-east-1:***:repository/deploy-airflow-on-ecs-fargate-airflow",
"registryId": "***",
"repositoryName": "deploy-airflow-on-ecs-fargate-airflow",
"repositoryUri": "***.dkr.ecr.us-east-1.amazonaws.com/deploy-airflow-on-ecs-fargate-airflow",
"createdAt": "2022-02-02T06:27:15+09:00",
"imageTagMutability": "MUTABLE",
"imageScanningConfiguration": {
"scanOnPush": true
},
"encryptionConfiguration": {
"encryptionType": "AES256"
}
}
]
}
- Authenticate your preferred container image build tool with AWS. The following works with Docker and Podman.
aws ecr get-login-password --region us-east-1 | docker login --username AWS --password-stdin ***.dkr.ecr.us-east-1.amazonaws.com
- Build and push the container image.
export REPO_URI="***.dkr.ecr.us-east-1.amazonaws.com/deploy-airflow-on-ecs-fargate-airflow"
docker buildx build -t "${REPO_URI}" -f build/prod/Containerfile --platform linux/amd64 .
docker push "${REPO_URI}"
- Deploy the remaining infrastructure.
terraform -chdir=infrastructure apply
- Initialize the airflow metadata database. Here we run the
db init
command as a standalone ECS task.
python3 scripts/run_task.py --wait-tasks-stopped --command 'db init'
- Create an admin user using the same method as
db init
.
python3 scripts/run_task.py --wait-tasks-stopped --command \
'users create --username airflow --firstname airflow --lastname airflow --password airflow --email airflow@example.com --role Admin'
- Find and open the airflow webserver load balancer URI.
aws elbv2 describe-load-balancers
{
"LoadBalancers": [
{
"DNSName": "airflow-webserver-231209530.us-east-1.elb.amazonaws.com",
(..redacted)
}
]
}
A common requirement is the ability to execute an arbitrary command in the cluster context. AWS provides the run-task API for this purpose.
The terraform code in this repository registers a template task definition named airflow-standalone-task
. To run arbitrary commands on the ECS cluster, override the default parameters in the task definition when calling run-task
. For example, one can run the command airflow db init
while specifying 1024 memory and 512 cpu.
python3 scripts/run_task.py --cpu 512 --memory 1024 --command 'db init'
This repository demonstrates various logging configurations.
Component | Log destination |
---|---|
Airflow Webserver, Scheduler & Metrics | CloudWatch |
Airflow Standalone Task | S3 via Kinesis Firehose (query-able with Athena) |
Airflow Worker | S3 via Airflow's builtin remote log handler |
A conservative estimate excluding free tier in which all ECS services run 24 hours a day (workers run at max capacity for 6 hours) costs around 200 USD per month. A similar configuration with Amazon Managed Workflows for Apache Airflow (MWAA) will cost at least 360 USD per month.
One can further limit costs by decreasing the max number of workers, or stopping the webserver and scheduler at night.
Airflow workers scale between 0-5 based on the current number of running, unpaused Airflow tasks. The Airflow statsd module does not provide this exact information, so I created a separate "metrics" ECS service to periodically query the metadata-db, compute the "desired worker count" using formulas described in the MWAA documentation and this AWS blog post, and send the custom metric data to CloudWatch. Metric data sent to CloudWatch is then used in a target tracking scaling policy to scale worker service containers.
The webserver and scheduler each scale to 1 in the morning and 0 at night using scheduled autoscaling. One could go a step further by using the ALBRequestCountPerTarget
predefined metric to scale the webserver via a target tracking scaling policy.
As described above, this repository registers a task definition named airflow-standalone-task
for the purpose of running one-off commands in the cluster context. Take a look inside scripts/run_task.py
to see how one can use the run-task API to override options like command
, cpu
and memory
when running a standalone task.
python3 scripts/run_task.py --command \
'users create --username airflow --firstname airflow --lastname airflow --password airflow --email airflow@example.com --role Admin'
Get a shell into a service container using ECS exec
One can use ECS exec to get a shell into a running container. Install the Session Manager plugin for awscli
, obtain the ID of the task container and execute the following command.
aws ecs execute-command --cluster airflow --task 9db18526dd8341169fbbe3e2b74547fb --container scheduler --interactive --command "/bin/bash"
If successful, a terminal prompt will appear.
The Session Manager plugin was installed successfully. Use the AWS CLI to start a session.
Starting session with SessionId: ecs-execute-command-0d94b5b2472323b7d
root@9db18526dd8341169fbbe3e2b74547fb-2568554522:/opt/airflow# ls -la
total 52
drwxrwxr-x 1 airflow root 4096 Feb 1 22:41 .
drwxr-xr-x 1 root root 4096 Jan 17 23:22 ..
-rw-r--r-- 1 airflow root 1854 Feb 1 22:37 airflow.cfg
drwxrwxr-x 1 airflow root 4096 Feb 1 21:28 dags
drwxr-xr-x 1 airflow root 4096 Feb 1 21:28 deploy_airflow_on_ecs_fargate
drwxrwxr-x 1 airflow root 4096 Feb 1 22:41 logs
drwxr-xr-x 2 airflow root 4096 Feb 1 21:28 plugins
-rw-r--r-- 1 airflow root 199 Jan 31 00:45 requirements.txt
-rw-rw-r-- 1 airflow root 4695 Feb 1 22:41 webserver_config.py
root@9db18526dd8341169fbbe3e2b74547fb-2568554522:/opt/airflow# whoami
root
ECS exec may fail to create a session and display the following message. If this happens, one can often mitigate the problem by forcing a new container deployment.
An error occurred (InvalidParameterException) when calling the ExecuteCommand operation: The execute command failed because execute command was not enabled when the task was run or the execute command agent isn’t running. Wait and try again or run a new task with execute command enabled and try again.
There are multiple ways to scale services. Here are some options using the commandline.
One can change the desired task count of a service via the update-service
API. Scaling actions take effect immediately.
aws ecs update-service --cluster airflow --service airflow-webserver --desired-count 0
If the service is registered as an autoscaling target (it is in this project), one can also set the desired count via a scheduled autoscaling action. This may be helpful if one wants to scale to N at a certain time.
# macos
export TWO_HOURS_LATER=$(date -u -v+2H '+%Y-%m-%dT%H:%M:00')
# linux
export TWO_HOURS_LATER=$(date -u --date='2 hours' '+%Y-%m-%dT%H:%M:00')
aws application-autoscaling put-scheduled-action \
--service-namespace ecs \
--scalable-dimension ecs:service:DesiredCount \
--resource-id service/airflow/airflow-webserver \
--scheduled-action-name scale-webserver-to-zero \
--schedule "at(${TWO_HOURS_LATER})" \
--scalable-target-action MinCapacity=0,MaxCapacity=0