-
Notifications
You must be signed in to change notification settings - Fork 14
Limit QuantumESPRESSO builds on A64FX to max 6 cores #106
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
bot: build repo:eessi.io-2023.06-software instance:eessi-bot-deucalion for:arch=aarch64/a64fx |
New job on instance
|
bot: build repo:eessi.io-2023.06-software instance:eessi-bot-deucalion for:arch=aarch64/a64fx |
New job on instance
|
Job is still running, but the QE build just completed. The max memory usage reported by Slurm is only 1953600K, so I don't understand why it didn't work with the default settings (which should be 12 cores instead of 6?). |
…-scripts into qe_numcores_a64fx
bot: build repo:eessi.io-2023.06-software instance:eessi-bot-deucalion for:arch=aarch64/a64fx |
New job on instance
|
New job on instance
|
Nodes have about 29G free memory for jobs. So if 6 use 19 G (6 * 2.5 G + 4 G), using 38 G for 12 cores (or say 12 * 2.5 G + 4 G = 34 g) would be too much. Anyhow, the true culprit has been found. |
True, but 1953600 K is only 1.9 GB 😉 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
staging PR merged |
And undo the limit for CP2K (introduced in #104), I don't think that was required and it didn't solve the issue in EESSI/software-layer#1220 (comment). Newer QE versions removed the
maxparallel=1
, and it looks like this makes it run out of memory.To be sure, I'll add an easystack here that builds both QE and CP2K, just to confirm that both build without issues now.