You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The current code appears to assume that each job uses a full node. This might not always be the case when using for example a slurm job array with multiple elements per node. This can lead to large over estimate of the cluster usage.
I'm guessing an (at least partial) fix would be to only count distinct nodes in the process_data function.
The text was updated successfully, but these errors were encountered:
I still think the right thing to do is to count all the cores on a node because in principal these are reserved and no other people can use them. However, like you suggest, we shouldn't double count nodes if multiple jobs of the same person run on a node. If multiple people run a job on a node it's more complicated though (and perhaps we can ignore this).
The current code appears to assume that each job uses a full node. This might not always be the case when using for example a slurm job array with multiple elements per node. This can lead to large over estimate of the cluster usage.
I'm guessing an (at least partial) fix would be to only count distinct nodes in the
process_data
function.The text was updated successfully, but these errors were encountered: