Open
Description
I am trying to use more than one GPU for warpx on our AMD machine without success.
The setting is the following
- AMD GPU MI 100 with ROCM 5.4.x
- warpx using the std [laser_ion](example https://github.com/ECP-WarpX/WarpX/blob/development/Examples/Physics_applications/laser_ion/inputs)
- submit slurm command:
sbatch --reservation gpu_tests --nodes 1 --ntasks-per-node 4 --cpus-per-task 1 --gres=gpu:4 --mem-per-gpu 48000 --no-requeue --job-name warpx --mail-type ALL --mail-user [email protected] --partition gpu --time 0-8:00:00 -D ./ -o %j.out.log -e %j.err.log --nodelist=lxbk1099 ./run-file.sh
From the ressource selection used (--gres=gpu:4
) i would expect the usage of 4 gpus. Instead only one is used .
Is there anything else one should be aware when running on multi gpus machine?
If yes could you deliver me an example i can test on our system ?