Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

MPI issues #25

Open
ghost opened this issue May 21, 2021 · 3 comments
Open

MPI issues #25

ghost opened this issue May 21, 2021 · 3 comments

Comments

@ghost
Copy link

ghost commented May 21, 2021

Hi!
Let me start by thanking you for your code!

We have implemented a new CV that calculates the cvvalue in parallel. That was needed as the algorithm considers the contribution of every atom in the system. We have a local version of SSAGES that contains the new piece of code. When everything works, we would like to share it with you. The cvvalue is calculated correctly across the processors. Unfortunately, we noticed that when running FFS with LAMMPS the “dumpfiles” are written by SSAGES serially, so each processor overwrites the output. So, e.g., in a file "l0-n0.dat" we only have information about the atoms handled by one core, and we receive the following error error, could not locate atomID 1 from dumpfile. We would like to use SSAGES for a large MD calculations. Therefore, parallelisation is important to us. Have we missed anything?

We would really appreciate your help.

@mquevill
Copy link
Collaborator

You are correct that multiple MPI processes are trying to read/write at the same time, which corrupts the dumpfiles. Because of this and a few other quirks, the current implementation of FFS is limited to just 1 MPI process per walker. We hope to fix the full parallelization capabilities of FFS soon, but we do not have an explicit timeline for when that will be. See also #10 (comment).

As a current workaround to accelerate simulations, you could try utilizing OpenMP threads with the USER-OMP package. It is only the MPI processes per walker that are limited.

@ghost
Copy link
Author

ghost commented May 25, 2021

Thank you for your response.

@mquevill
Copy link
Collaborator

SSAGES v0.9.3 will now throw an error if the user tries to use more than 1 MPI process per walker. While we would like to get full MPI support for FFS, this should prevent the generation of corrupted data files.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant