Skip to content

Apache Airflow Add-on

Apache Airflow is an open-source platform for developing, scheduling, and monitoring batch-oriented workflows. Airflow contains extensible framework that allows for building workflows connecting with many technologies. For more information on Airflow, please consult the official documentation.

This add-on is an implementation of Apache Airflow on EKS using the official helm chart.


  1. If you are using an S3 bucket (for logging storage) or EFS File System (for DAGs persistent storage), you must register the resources using Resource Provider. The framework provides S3 and EFS Resource Provider.
  2. If you are using an EFS File System, you must include an EFS CSI Driver in the add-on array, otherwise will run into the following error: Missing a dependency: EfsCsiDriverAddOn. Please add it to your list of addons..
  3. If you are using an Ingress with AWS Load Balancer, you must include an AWS Load Balacner Controller in the add-on array, otherwise you will run into the following error: Missing a dependency: AwsLoadBalancerControllerAddOn. Please add it to your list of addons.


import * as cdk from 'aws-cdk-lib';
import * as blueprints from '@aws-quickstart/eks-blueprints';

const app = new cdk.App();

const apacheAirflowS3Bucket = new blueprints.CreateS3BucketProvider({
    id: 'apache-airflow-s3-bucket-id',
    s3BucketProps: { removalPolicy: cdk.RemovalPolicy.DESTROY }
const apacheAirflowEfs = new blueprints.CreateEfsFileSystemProvider({
    name: 'blueprints-apache-airflow-efs',    

const addOn = [new blueprints.EfsCsiDriverAddOn(),
    new blueprints.ApacheAirflowAddOn({
        enableLogging: true,
        s3Bucket: 'airflow-logging-s3-bucket',
        enableEfs: true,
        efsFileSystem: 'apache-airflow-efs-provider',

const blueprint = blueprints.EksBlueprint.builder()
  .resourceProvider('apache-airflow-s3-bucket-provider', apacheAirflowS3Bucket)
  .resourceProvider('apache-airflow-efs-provider', apacheAirflowEfs)
  .build(app, 'my-stack-name');


To validate that the Airflow add-on is installed properly, ensure that the required kubernetes resources are running in the cluster:

kubectl get all -n airflow
NAME                                                             READY   STATUS    RESTARTS     AGE
pod/blueprints-addon-apache-airflow-postgresql-0                 1/1     Running   0            1m4s
pod/blueprints-addon-apache-airflow-scheduler-697958497d-tbblk   2/2     Running   0            1m4s
pod/blueprints-addon-apache-airflow-statsd-5b97b9fcb4-9r8qc      1/1     Running   0            1m4s
pod/blueprints-addon-apache-airflow-triggerer-86b94646c6-xrjnn   2/2     Running   0            1m4s
pod/blueprints-addon-apache-airflow-webserver-6b8db695fc-9d87q   1/1     Running   0            1m4s

NAME                                                    TYPE        CLUSTER-IP       EXTERNAL-IP   PORT(S)             AGE
service/blueprints-addon-apache-airflow-postgresql      ClusterIP    <none>        5432/TCP            1m4s
service/blueprints-addon-apache-airflow-postgresql-hl   ClusterIP   None             <none>        5432/TCP            1m4s
service/blueprints-addon-apache-airflow-statsd          ClusterIP   <none>        9125/UDP,9102/TCP   1m4s
service/blueprints-addon-apache-airflow-webserver       ClusterIP    <none>        8080/TCP            1m4s

NAME                                                        READY   UP-TO-DATE   AVAILABLE   AGE
deployment.apps/blueprints-addon-apache-airflow-scheduler   1/1     1            1           1m4s
deployment.apps/blueprints-addon-apache-airflow-statsd      1/1     1            1           1m4s
deployment.apps/blueprints-addon-apache-airflow-triggerer   1/1     1            1           1m4s
deployment.apps/blueprints-addon-apache-airflow-webserver   1/1     1            1           1m4s

NAME                                                                   DESIRED   CURRENT   READY   AGE
replicaset.apps/blueprints-addon-apache-airflow-scheduler-669dff9c6d   0         0         0       1m4s
replicaset.apps/blueprints-addon-apache-airflow-scheduler-697958497d   1         1         1       1m4s
replicaset.apps/blueprints-addon-apache-airflow-statsd-5b97b9fcb4      1         1         1       1m4s
replicaset.apps/blueprints-addon-apache-airflow-triggerer-7b7c69486d   0         0         0       1m4s
replicaset.apps/blueprints-addon-apache-airflow-triggerer-86b94646c6   1         1         1       1m4s
replicaset.apps/blueprints-addon-apache-airflow-webserver-5db49dcb94   0         0         0       1m4s
replicaset.apps/blueprints-addon-apache-airflow-webserver-6b8db695fc   1         1         1       1m4s

NAME                                                          READY   AGE
statefulset.apps/blueprints-addon-apache-airflow-postgresql   1/1     1m4s


To test the Airflow functionality, expose the webserver by port forwarding:

kubectl port-forward service/blueprints-addon-apache-airflow-webserver -n airflow 8080:8080

You should be able to access the webserver via your browser: http://localhost:8080. Default username and password are both admin.


Applies the Apache Airflow add-on to an Amazon EKS cluster. Optionally, you can leverage integrations with the following AWS services out of the box:

  1. S3 Bucket for storing server, worker, and scheduler logs remotely.
  2. EFS File System for storing DAGs
  3. AWS Load Balancer for ingress to the webserver
  4. A Certificate for AWS Certificate Manager for SSL