-
Notifications
You must be signed in to change notification settings - Fork 7
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[0.25deg] update PE layout #214
Comments
You'll also need to update the |
Right, thanks @dougiesquire |
I can't remember where we got those block sizes from, we should get better performance if we can reduce max_blocks (say to 10?) by setting the blocksizes differently. Sorry I was wrong last week - we did put in a patch for max_blocks ... you can remove it from the namelist. Its still good to check the logs to get it to closer to 10. The process would be - pick number of procs, then set block_size_x & block_size_y such that the blocks are close to square, and there are around 10 per PE (ideally also nx_global is divisible by block_size_x and ny_global is divisible by block_size_y) We can also remove debug_blocks - but whilst setting the block size it provides useful information |
I came across this issue again #156, where I forgot to adjust the
@anton-seaice I understand the calculations, but could you please clarify why the ICE The error message isn’t very intuitive, making it difficult for users to realise that they need to modify these parameters when changing the layout. Can we revert it to the settings used in the 1deg configuration, here https://github.com/ACCESS-NRI/access-om3-configs/blob/2bc6107ef1b195aa62485a5d87c4ba834996d8cc/nuopc.runconfig#L364-L373?
|
The block sizes were adopted from OM2 report, which specifies a CICE5 block size of
I cant remember why having the number of blocks close to 10? |
In the old COSIMA TWG minutes from OM2 development (on the COSIMA website) the recommendation from NCI was to use one task per node. I think the Yang 2019 on Parallel I/O in MOM5 makes similar suggestion ? I guess there is a hardware benefit to one task per node. There's so many options there its hard to know what the best combination is without lots of work. e.g. we could also test having a dedicated IO PE, or changing the PIO_rearranger I think one IO task per node is a good start. We could try just one IO task, it might not make much difference at this resolution.
I agree, does it make a seperate ESMF log file ? I think they have names something like
Ok thanks!
From the cice docs : Smaller, more numerous blocks provides an opportunity for better load balance by allocating each processor both ice-covered and ice-free blocks. But smaller, more numerous blocks becomes less efficient due to MPI communication associated with halo updates. In practice, blocks should probably not have fewer than about 8 to 10 grid cells in each direction, and more square blocks tend to optimize the volume-to-surface ratio important for communication cost. Often 3 to 8 blocks per processor provide the decompositions flexiblity to create reasonable load balance configurations. So we should actually aim for number of blocks of 8 or less by the sounds of it :) |
I agree for the current phase. I will do a test on the I/O tasks to verify the optimal configuration.
This can be enabled by setting this
The updated settings result in a
|
When setting
Despite this warning, I don’t believe it will impact overall performance since MOM typically has a much higher computational load than CICE. NB: |
Why do you think max_blocks shouldn't be 5 ? |
It can be 5, but we have to manually modify it to be 5 |
Oh sorry, I see now. That's something about the patch we put into access-om3 0.3.x for removing max_blocks, and the max_blocks calculation being approximate. When we update the cice version it should go away (after CICE-Consortium/CICE#954) It will allocate ~20% more memory than it uses , but it uses a small enough amount of memory there probably isn't a performance impact. |
I created payu-org/payu#496 to add checks for the iolayout numbers |
Closed through ACCESS-NRI/access-om3-configs#114 |
To update the PE layout for the 0.25 deg configuration,
nuopc.runconfig
,ice_in
andconfig.yaml
require corresponding modifications.Note: These changes will be updated when the configuration is revised.
nuopc.runconfig
ice_in
config.yaml
The text was updated successfully, but these errors were encountered: