File tree Expand file tree Collapse file tree 6 files changed +13
-13
lines changed Expand file tree Collapse file tree 6 files changed +13
-13
lines changed Original file line number Diff line number Diff line change @@ -26,8 +26,7 @@ INPUTS=inputs
2626# per node are 2x 2.4 GHz Intel Xeon Gold 6148
2727# note: the system seems to only expose cores (20 per socket),
2828# not hyperthreads (40 per socket)
29- export SRUN_CPUS_PER_TASK=20
30- export OMP_NUM_THREADS=${SRUN_CPUS_PER_TASK}
29+ export OMP_NUM_THREADS=${SLURM_CPUS_PER_TASK}
3130
3231# GPU-aware MPI optimizations
3332GPU_AWARE_MPI=" amrex.use_gpu_aware_mpi=1"
Original file line number Diff line number Diff line change 2525# SBATCH -o stdout_%j
2626# SBATCH -e stderr_%j
2727
28- # OpenMP threads per MPI rank
29- export OMP_NUM_THREADS=16
30- export SRUN_CPUS_PER_TASK=16
31-
3228# set user rights to u=rwx;g=r-x;o=---
3329umask 0027
3430
31+ # OpenMP threads per MPI rank
32+ export OMP_NUM_THREADS=${SLURM_CPUS_PER_TASK}
33+
3534# executable & inputs file or python interpreter & PICMI script here
3635EXE=./warpx.rz
3736INPUTS=./inputs_rz
Original file line number Diff line number Diff line change 1414# SBATCH -q regular
1515# SBATCH -C gpu
1616# SBATCH --exclusive
17+ # SBATCH --cpus-per-task=32
1718# SBATCH --gpu-bind=none
1819# SBATCH --gpus-per-node=4
1920# SBATCH -o WarpX.o%j
@@ -27,7 +28,7 @@ INPUTS=inputs_small
2728export MPICH_OFI_NIC_POLICY=GPU
2829
2930# threads for OpenMP and threaded compressors per MPI rank
30- export SRUN_CPUS_PER_TASK=32
31+ export OMP_NUM_THREADS= ${SLURM_CPUS_PER_TASK}
3132
3233# depends on https://github.com/ECP-WarpX/WarpX/issues/2009
3334# GPU_AWARE_MPI="amrex.the_arena_is_managed=0 amrex.use_gpu_aware_mpi=1"
Original file line number Diff line number Diff line change 1313# SBATCH -A <proj>
1414# SBATCH -q regular
1515# SBATCH -C cpu
16+ # 8 cores per chiplet, 2x SMP
17+ # SBATCH --cpus-per-task=16
1618# SBATCH --ntasks-per-node=16
1719# SBATCH --exclusive
1820# SBATCH -o WarpX.o%j
@@ -30,10 +32,9 @@ INPUTS=inputs_small
3032# This will be our MPI rank assignment (2x8 is 16 ranks/node).
3133
3234# threads for OpenMP and threaded compressors per MPI rank
33- export SRUN_CPUS_PER_TASK=16 # 8 cores per chiplet, 2x SMP
3435export OMP_PLACES=threads
3536export OMP_PROC_BIND=spread
36- export OMP_NUM_THREADS=${SRUN_CPUS_PER_TASK }
37+ export OMP_NUM_THREADS=${SLURM_CPUS_PER_TASK }
3738
3839srun --cpu-bind=cores \
3940 ${EXE} ${INPUTS} \
Original file line number Diff line number Diff line change 1717# A100 80GB (256 nodes)
1818# S BATCH -C gpu&hbm80g
1919# SBATCH --exclusive
20+ # SBATCH --cpus-per-task=16
2021# ideally single:1, but NERSC cgroups issue
2122# SBATCH --gpu-bind=none
2223# SBATCH --ntasks-per-node=4
@@ -33,8 +34,7 @@ export MPICH_OFI_NIC_POLICY=GPU
3334
3435# threads for OpenMP and threaded compressors per MPI rank
3536# note: 16 avoids hyperthreading (32 virtual cores, 16 physical)
36- export SRUN_CPUS_PER_TASK=16
37- export OMP_NUM_THREADS=${SRUN_CPUS_PER_TASK}
37+ export OMP_NUM_THREADS=${SLURM_CPUS_PER_TASK}
3838
3939# GPU-aware MPI optimizations
4040GPU_AWARE_MPI=" amrex.use_gpu_aware_mpi=1"
Original file line number Diff line number Diff line change 1212# SBATCH -J WarpX
1313# S BATCH -A <proj> # project name not needed yet
1414# SBATCH -p mi300a
15+ # SBATCH --cpus-per-task=16
1516# SBATCH --gpu-bind=none
1617# SBATCH --ntasks-per-node=4
1718# SBATCH --gpus-per-node=4
@@ -27,8 +28,7 @@ export MPICH_OFI_NIC_POLICY=GPU
2728
2829# threads for OpenMP and threaded compressors per MPI rank
2930# note: 16 avoids hyperthreading (32 virtual cores, 16 physical)
30- export SRUN_CPUS_PER_TASK=16
31- export OMP_NUM_THREADS=${SRUN_CPUS_PER_TASK}
31+ export OMP_NUM_THREADS=${SLURM_CPUS_PER_TASK}
3232
3333# GPU-aware MPI optimizations
3434GPU_AWARE_MPI=" amrex.use_gpu_aware_mpi=1"
You can’t perform that action at this time.
0 commit comments