This repository contains the code for running a high-performance computing (HPC) pipeline that get the mosquito basic reproductive number based on https://github.com/mpardo1/RM_mosquito
The pipeline is orchestrated using Snakemake which automates tasks related to data preparation, model training, and prediction.
The model needs data from the following data sources:
- Realtime ECMWF - HRES
- GPWv4 - Population Density
In the case of future forecast, please consider (not yet tested):
- CMIP6 climate projections
- GHS-POP for future estimates of human population density.
Please download the file gpw-v4-population-density-rev11_2020_2pt5_min_tif.zip export it and place it in the following path: data/gpw/gpw_v4_population_density_rev11_2020_2pt5_min.tif
We recommend using Mamba (a faster drop-in replacement for Conda). If you don't have Conda or Mamba installed, consider installing Miniforge.
Install Snakemake, Snakedeploy, and necessary plugins:
mamba create -c conda-forge -c bioconda --name snakemake snakemake=9.8.1 snakedeploy=0.11.0If you're running on an HPC with SLURM, install additional plugins:
mamba install -n snakemake -c bioconda snakemake-executor-plugin-slurm=1.5.0Activate the environment:
conda activate snakemakeCreate and move into a project directory:
mkdir -p path/to/project-workdir
cd path/to/project-workdirDeploy the workflow using Snakedeploy:
snakedeploy deploy-workflow <URL_TO_THIS_REPO> . --tag <DESIRED_TAG>This will create two directories:
workflow/: contains the deployed Snakemake moduleconfig/: contains configuration files
Edit config/config.yaml to specify your settings (paths, parameters, etc.) according to your data and environment
snakemake --cores all --sdm condaUse the provided SLURM profile:
snakemake --cores all --sdm conda --profile slurmFor advanced features such as cluster execution, cloud deployments, and workflow customization, see the Snakemake documentation.