-
Notifications
You must be signed in to change notification settings - Fork 26
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Default GCHP run crashes almost immediately in MAPL_CapGridComp.F90 #8
Comments
Thanks @LiamBindle. Do you know if using mpiuni set mpi to automatically use one core per node? Does this implementation exist for some specific purpose? |
@lizziel According to this mpiuni is "a single-processor MPI-bypass library". I don't really understand what that means, but it seems to fit inline with what I was seeing (each process thinking they were root). Actually, last night I just noticed ESMF_COMM wasn't "intelmpi", and I thought it should be, so I rebuilt ESMF and it fixed my problem. I tried to generalize this issue for the purpose of the issue archive. This issue is really "ESMF was built with the wrong ESMF_COMM" and isn't because of a ESMF + Intel MPI compatibility problem. Here is the part of spack that sets ESMF_COMM. It looks like |
This merge brings PR #395 (Expand GCHP advection diagnostics, by @lizziel) into the GCHP "no-diff-to-benchmark" development stream. This PR contains the changelog updates for GCHP advection diagnostic updates going into FV3 and GEOS-Chem submodules. It also fixes a bug in setting the UpwardsMassFlux flux diagnostic that was causing it to be all zeros. This PR should be merged in at the same time geos-chem PR #2199 and FVdycoreCubed_GridComp #8. Signed-off-by: Bob Yantosca <yantosca@seas.harvard.edu>
This commit informs the GCHP superproject about the following commits that were pushed to the GitHub geoschem/geos-chem repository: af42462 Merge PR #8 (Add PLEadv diagnostic for offline advection in GCHP) This PR adds the PLEadv diagnostic export to FvDyCore_Cubed. Signed-off-by: Bob Yantosca <yantosca@seas.harvard.edu>
Hi everyone,
I'm just submitting this for the archive of issues on GitHub.
Relevent Information
ESMF_COMM=mpiuni
What happened
Yesterday I tried running the default 6-core 1-node 1-hour GCHP simulation and it crashed almost immediately. This happned with GHCP_CTM 13.0.0-alpha.1, but this could happen with any version that uses MAPL 2.0+. Below is the full output. The important parts to pick out are:
Failed run output:
The Problem
The issue was ESMF was built with
ESMF_COMM=mpiuni
. This appears to have happended because the spack install spec wasn't quite right, but I didn't build ESMF myself so I can't be sure.How do I check which ESMF_COMM my ESMF was built with?
The build-time value of ESMF_COMM is written to
esmf.mk
beside your ESMF libraries. You can see it with the following commandor
Solution
Rebuild ESMF and make sure
ESMF_COMM
is set to the appropriate MPI flavor.The text was updated successfully, but these errors were encountered: