I tried starting grappa's bfs implementation on 4 nodes, each equipped with 64GB of ram and one Intel Xeon CPU with 6 cores (two threads per core). I used slurm to start the job and want to load a scale 26 graph which was generated by the graph500 graph generator (total file size ~16GB).
Running it with "srun --nodes=4 --ntasks-per-node=12 -- ./bfs_beamer.exe --path=/some/path/rmat_26_16" throws the following error:
"Out of memory in the global heap: couldn't find a chunk of size 34359738368 to hold an allocation of 34359738368 bytes. Can you increase --global_heap_fraction?"
I changed the --global_heap_fraction to various values (0.1 to 1.0) but none of them worked. I have found your notes on setting the memory sizes (https://github.com/uwsampa/grappa/wiki/Setting-memory-region-sizes) but that didn't help me figuring out any working combination of settings.
Are there any other settings I have to change or have i hit a limit with the given setup? For evaluation purpose, I have to find out what's the biggest synthetic graph I can load with the given nodes using grappa.
Full output of the srun command without --global_heap_fraction option set:
http://pastebin.com/6GrCbPdN
I tried starting grappa's bfs implementation on 4 nodes, each equipped with 64GB of ram and one Intel Xeon CPU with 6 cores (two threads per core). I used slurm to start the job and want to load a scale 26 graph which was generated by the graph500 graph generator (total file size ~16GB).
Running it with "srun --nodes=4 --ntasks-per-node=12 -- ./bfs_beamer.exe --path=/some/path/rmat_26_16" throws the following error:
"Out of memory in the global heap: couldn't find a chunk of size 34359738368 to hold an allocation of 34359738368 bytes. Can you increase --global_heap_fraction?"
I changed the --global_heap_fraction to various values (0.1 to 1.0) but none of them worked. I have found your notes on setting the memory sizes (https://github.com/uwsampa/grappa/wiki/Setting-memory-region-sizes) but that didn't help me figuring out any working combination of settings.
Are there any other settings I have to change or have i hit a limit with the given setup? For evaluation purpose, I have to find out what's the biggest synthetic graph I can load with the given nodes using grappa.
Full output of the srun command without --global_heap_fraction option set:
http://pastebin.com/6GrCbPdN