You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This is an AWS Executor that delegates every task to a scheduled container on either AWS Batch, AWS Fargate, or AWS ECS.
5
7
@@ -12,37 +14,6 @@ For `AWS Batch`: [Getting Started with AWS Batch ReadMe](getting_started_batch.m
12
14
13
15
For `AWS ECS/Fargate`: [Getting Started with AWS ECS/Fargate ReadMe](getting_started_ecs_fargate.md)
14
16
15
-
16
-
## How Airflow Executors Work
17
-
Every time Apache Airflow wants to run a task, the Scheduler generates a shell command that needs to be executed **somewhere**.
18
-
Under the hood this command will run Python code, and it looks something like this:
19
-
```bash
20
-
airflow run <DAG_ID><TASK_ID><EXECUTION_DATE>
21
-
```
22
-
What people refer to as the Executor, is actually an API and not a thread/process. The process that runs is the Airflow
23
-
Scheduler, and the Executor API is part of that process. When the Scheduler process generates the shell command above,
24
-
it is the executor that decides how this is ran. Here's how what different Executors handle this command:
25
-
***LocalExecutor**
26
-
*`execute_async()` Uses Python's Subprocess library to spin up a new process with the shell command
27
-
*`sync()` Monitors exit code on every heartbeat
28
-
***CeleryExecutor**
29
-
*`execute_async()` Uses the Celery library to put this shell command in a message queue
30
-
*`sync()` Monitors the Celery Backend to determine task completion
31
-
***KubernetesExecutor**
32
-
*`execute_async()` Launches K8 Pod with the shell command
33
-
*`sync()` Monitors K8 Pod
34
-
35
-
So, how do these executors work? Well, on the highest level, it just calls Boto3 APIs to schedule containers onto
36
-
compute clusters.
37
-
38
-
***AwsEcsFargateExecutor**
39
-
*`execute_async()` Uses the Boto3 library to call [ECS.run_task()][run_task], and launches a container with the shell command onto a EC2 or Serverless cluster
40
-
*`sync()` Uses the Boto3 library to call ECS.describe_tasks() to monitor the exit-code of the Airflow Container.
41
-
***AwsBatchExecutor**
42
-
*`executor_async()`Uses the Boto3 library to call [Batch.submit_job()][submit_job], and puts this message in a job-queue
43
-
*`sync()` Uses the Boto3 library to call Batch.describe_jobs() to monitor the status of the Airflow Container
44
-
45
-
46
17
## But Why?
47
18
There's so much to unpack here.
48
19
@@ -271,7 +242,7 @@ Please file a ticket in GitHub for issues. Be persistent and be polite.
271
242
272
243
273
244
## Contribution & Development
274
-
This repository uses Travis-CI for CI, pytest for Integration/Unit tests, and isort+pylint for code-style.
245
+
This repository uses Github Actions for CI, pytest for Integration/Unit tests, and isort+pylint for code-style.
275
246
Pythonic Type-Hinting is encouraged. From the bottom of my heart, thank you to everyone who has contributed
0 commit comments