|
24 | 24 | DEFAULT_THREADS, |
25 | 25 | ) |
26 | 26 | from generate.ghz import gen_circuit |
| 27 | +from export.kaggle import upload_dataset as upload_dataset_kaggle |
| 28 | +from export.huggingface import upload_dataset as upload_dataset_huggingface |
27 | 29 |
|
28 | 30 | default_args = { |
29 | 31 | "depends_on_past": True, |
|
92 | 94 | with resized and normalized images. |
93 | 95 | """ |
94 | 96 |
|
95 | | - command = f"zip -r {folder}/dataset-images.zip {folder}/dataset/" |
| 97 | + command = f"cd {folder} && zip -r dataset-images.zip dataset/" |
96 | 98 | pack_img = BashOperator(task_id="pack_images", bash_command=command) |
97 | 99 |
|
98 | 100 | pack_img.doc_md = """ |
|
118 | 120 | Run training after finishing all processes. |
119 | 121 | """ |
120 | 122 |
|
| 123 | + kaggle_dataset = os.environ.get("KAGGLE_DATASET") |
| 124 | + send_kaggle = PythonOperator( |
| 125 | + task_id="send_kaggle", |
| 126 | + python_callable=upload_dataset_kaggle, |
| 127 | + op_args=[kaggle_dataset,folder] |
| 128 | + ) |
| 129 | + |
| 130 | + hf_dataset = os.environ.get("HF_DATASET") |
| 131 | + send_hf = PythonOperator( |
| 132 | + task_id="send_huggingface", |
| 133 | + python_callable=upload_dataset_huggingface, |
| 134 | + op_args=[hf_dataset,folder] |
| 135 | + ) |
| 136 | + |
| 137 | + send_hf.doc_md = """ |
| 138 | + Send dataset files to huggingface |
| 139 | + """ |
| 140 | + |
121 | 141 | create_folder >> [gen_ghz, gen_df] |
122 | 142 | gen_df >> gen_images |
123 | 143 | gen_images >> remove_duplicates |
124 | 144 | remove_duplicates >> transform_img |
125 | 145 | transform_img >> pack_img |
126 | 146 |
|
127 | 147 | [gen_ghz, pack_img] >> trigger_dag_train |
| 148 | + [gen_ghz, pack_img] >> send_kaggle |
| 149 | + [gen_ghz, pack_img] >> send_hf |
| 150 | + |
0 commit comments