-
Notifications
You must be signed in to change notification settings - Fork 1
Improved config #1
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: 8km_jra_ryf_obc2-sapphirerapid-Charrassin-newparams-rerun-Wright-spinup-accessom2IC
Are you sure you want to change the base?
Improved config #1
Conversation
9a0f58d to
a023cb1
Compare
|
Thanks so much @edoyango this is awesome! Since I needed to modify my run a little after 8 years due to a salt restoring file bug I found, I decided to swap to SR with your current optimisation for the last 2 years of my spinup. I merged your commits in this PR into my spinup config (which I'd previously done on cascade lake) in this branch https://github.com/claireyung/access-om3-configs/tree/8km_jra_ryf_obc2-sapphirerapid-Charrassin-newparams-rerun-Wright-spinup-accessom2IC-yr9 The cost before (cascade lake, Helen's executable, Is this the kind of speed that was expected? Note 2603/10days x 31days ~ 8100 SU * Naively, I guess the netcdf files are bigger at the end when you run for a month vs 10 days which maybe slows down the final steps of the model and makes it slightly more than the raw scaling? *this is not really a fair comparison, because the config I gave you has ^all quoted numbers compare Januarys, but I haven't looked at the range |
Yes I think this is roughly what I would expect. Some of the changes, especially the IO_layout improvements, primarily affect the final dumping of restart files (which take a lot of time). Since your runs are much longer, you won't see as much of a benefit from those.
This is an interesting change in the config! This will probably affect who's waiting for who. I'll do more testing - the PE_LAYOUT from before was assuming MOM was the bottleneck, but now that CICE has more dynamic steps, giving more cores to CICE might be beneficial? |
|
Hi @claireyung, just following up on this
Just to say that I couldn't get a conclusive answer on this - increasing/decreasing cores assigned to MOM didn't seem to change much. |
|
Hey @edoyango, thanks for looking at it and for the update! I guess maybe this means we are approaching an optimum model cost... |
Hi @claireyung,
This is a pull request which tracks some of the config improvements that you could leverage (on top of the improvements to the exe done here). I'll populate this description as results are found.
PARALLEL_RESTART = True, but that doesn't work withpayuMinghang's looking at it. Is answer changingnuopc.runseq[x, y] = [60, 54]) SU seemed to increase a tiny bit, but reported CICE time (fromice.logsays CICE walltime went down by 5% to 478.34). For more ice scenarios, having larger block sizes seemed to improve performance significantly. See zulip comment for more details.sectrobindistributiondistribution_type=sectrobinis a bit better for PE layout in terms of neighbours/communication, but a bit worse for load balancing. This reduced CICE time (as perice.logto 449.34s walltime).