@@ -48,38 +48,48 @@ Build this image and upload it to a private repository. You may want to use Dock
48
48
the container. * BE SURE YOU HAVE THIS * ! It' s heavily important that commands like
49
49
`[" airflow" , " run" , < dag_id> , < task_id> , < execution_date> ]` are accepted by your container' s entrypoint script.
50
50
51
- 4 . Run through the AWS Batch Creation Wizard on AWS Console. The executor does not have any
51
+ 4 . Run through the AWS Batch Creation Wizard on AWS Console. The executor does not have any
52
52
prerequisites to how you create your Job Queue or Compute Environment. Go nuts; have at it. I' ll refer you to the
53
53
[AWS Docs' Getting Started with Batch](https://docs.aws.amazon.com/batch/latest/userguide/Batch_GetStarted.html).
54
54
You will need to assign the right IAM roles for the remote S3 logging.
55
55
Also, your dynamically provisioned EC2 instances do not need to be connected to the public internet,
56
56
private subnets in private VPCs are encouraged. However, be sure that all instances has access to your Airflow MetaDB.
57
57
58
- 5 . When creating a Job Definition choose the ' container' type and point to the private repo. The ' commands' array is
58
+ 5 . When creating a Job Definition choose the ' container' type and point to the private repo. The ' commands' array is
59
59
optional on the task- definition level. At runtime, Airflow commands will be injected here by the AwsBatchExecutor!
60
60
61
- 6 . Let' s go back to that machine in step #1 that' s running the Scheduler. We' ll use the same docker container as
61
+ 6 . Let' s go back to that machine in step #1 that' s running the Scheduler. We' ll use the same docker container as
62
62
before; except we' ll do something like `docker run ... airflow webserver` and `docker run ... airflow scheduler`.
63
63
Here are the minimum IAM roles that the executor needs to launch tasks, feel free to tighten the resources around the
64
64
job- queues and compute environments that you' ll use.
65
- ```json
66
- {
67
- " Version" : " 2012-10-17" ,
68
- " Statement" : [
69
- {
70
- " Sid" : " AirflowBatchRole" ,
71
- " Effect" : " Allow" ,
72
- " Action" : [
73
- " batch:SubmitJob" ,
74
- " batch:DescribeJobs" ,
75
- " batch:TerminateJob"
76
- ],
77
- " Resource" : " *"
78
- }
79
- ]
80
- }
81
- ```
82
- 7 . You' re done. Configure & launch your scheduler. However, maybe you did something real funky with your AWS Batch compute
65
+ ```json
66
+ {
67
+ " Version" : " 2012-10-17" ,
68
+ " Statement" : [
69
+ {
70
+ " Sid" : " AirflowBatchRole" ,
71
+ " Effect" : " Allow" ,
72
+ " Action" : [
73
+ " batch:SubmitJob" ,
74
+ " batch:DescribeJobs" ,
75
+ " batch:TerminateJob"
76
+ ],
77
+ " Resource" : " *"
78
+ }
79
+ ]
80
+ }
81
+ ```
82
+ 7 . Let' s test your AMI configuration before we launch the webserver or scheduler.
83
+ [Copy & Run this python file somewhere on the machine that has your scheduler](batch_ami_helper.py).
84
+ To simplify, this file will make sure that your scheduler and your cluster have correct AMI permissions. Your scheduler
85
+ should be able to launch, describe, plus terminate Batch jobs, and connect to your Airflow Meta DB of choice.
86
+ Meanwhile, your cluster should have the ability to pull the docker container, and also
87
+ connect to your Airflow MetaDB of choice.
88
+ ```bash
89
+ python3 - m unittest batch_ami_helper.py
90
+ ```
91
+
92
+ 9 . You' re done. Configure & launch your scheduler. However, maybe you did something real funky with your AWS Batch compute
83
93
environment. The good news is that you have full control over how the executor submits jobs.
84
94
See the [# Extensibility](./readme.md) section in the readme. Thank you for taking the time to set this up!
85
95
0 commit comments