Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
14 changes: 7 additions & 7 deletions scripts/run_aws_pipeline.sh
Original file line number Diff line number Diff line change
Expand Up @@ -250,15 +250,15 @@ fi
exit_code=$?

if [ $exit_code -eq 100 ]; then
echo "Success: Early exit, products already exist for all bursts."
echo "Success (100): Early exit, products already exist for all bursts."
exit 0 # Graceful exit with success code 0
fi
if [ $exit_code -eq 101 ]; then
echo "Process failed (101): Static Layers are missing."
exit 101
fi
if [ $exit_code -ne 0 ]; then
echo "Process failed: get-data-for-scene-and-make-run-config"
echo "Process failed (1): get-data-for-scene-and-make-run-config"
exit 1
fi

Expand All @@ -268,8 +268,8 @@ conda activate RTC
rtc_s1.py $RUN_CONFIG_PATH

if [ $? -ne 0 ]; then
echo "Process failed: rtc_s1.py $RUN_CONFIG_PATH"
exit 1
echo "Process failed (2): rtc_s1.py $RUN_CONFIG_PATH"
exit 2
fi

## -- MAKE THE METADATA FOR PRODUCTS AND UPLOAD TO S3 --
Expand Down Expand Up @@ -310,9 +310,9 @@ fi
exit_code=$?

if [ $exit_code -ne 0 ]; then
echo "Process failed: make-rtc-opera-stac-and-upload-bursts"
exit 1
echo "Process failed (3): make-rtc-opera-stac-and-upload-bursts"
exit 3
else
echo "Success: required burst products created."
echo "Success (0): required burst products created!"
exit 0
fi
10 changes: 7 additions & 3 deletions tests/sar_pipeline/test_full_aws_docker_run.py
Original file line number Diff line number Diff line change
Expand Up @@ -33,6 +33,7 @@
logger = logging.getLogger(__name__)

CURRENT_DIR = Path(__file__).parent.resolve()
TEST_OUTPUTS_DIR = f"{CURRENT_DIR}/data/isce3_rtc/results"
PROJECT_ROOT = CURRENT_DIR.parent.parent

REQUIRED_ENV_VARIABLES = [
Expand Down Expand Up @@ -122,19 +123,22 @@ def test_docker_with_args():
test_name = Path(__file__).stem
test_s3_bucket = "deant-data-public-dev"
test_s3_project_folder = f"TMP/sar-pipeline/{run_dt}/{test_name}"
if not Path(TEST_OUTPUTS_DIR).exists():
os.makedirs(TEST_OUTPUTS_DIR)
logging.info(f"Saving test outputs locally to : {TEST_OUTPUTS_DIR}")
logging.info(f"Uploading outputs to : {test_s3_bucket}/{test_s3_project_folder}")
docker_tag = re.sub(r"[^a-zA-Z0-9_.-]", "-", sar_pipeline.__version__)

logging.info(f"RUN 1: Producing Static Layers")
logging.info(
"Mounting test data directory for results: "
f"{CURRENT_DIR}/data/isce3_rtc/results:/home/rtc_user/working/results",
f"{TEST_OUTPUTS_DIR}:/home/rtc_user/working/results",
)
cmd = [
"docker",
"run",
"-v",
f"{CURRENT_DIR}/data/isce3_rtc/results:/home/rtc_user/working/results",
f"{TEST_OUTPUTS_DIR}:/home/rtc_user/working/results",
"--rm",
*ENV_VARS,
f"sar-pipeline:{docker_tag}",
Expand Down Expand Up @@ -168,7 +172,7 @@ def test_docker_with_args():
"docker",
"run",
"-v",
f"{CURRENT_DIR}/data/isce3_rtc/results:/home/rtc_user/working/results",
f"{TEST_OUTPUTS_DIR}:/home/rtc_user/working/results",
"--rm",
*ENV_VARS,
f"sar-pipeline:{docker_tag}",
Expand Down
Loading