Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Use dask cluster in analyze_ensemble.py #48

Open
FariborzDaneshvar-NOAA opened this issue Apr 11, 2024 · 1 comment
Open

Use dask cluster in analyze_ensemble.py #48

FariborzDaneshvar-NOAA opened this issue Apr 11, 2024 · 1 comment
Assignees

Comments

@FariborzDaneshvar-NOAA
Copy link
Collaborator

FariborzDaneshvar-NOAA commented Apr 11, 2024

@SorooshMani-NOAA as we discussed before, analyze_ensemble.py might not successfully run for big storms/large ensemble due to memory issue.
A quick fix, as you suggested, was adding dask cluster under the if __name__ == '__main__': and before _analyze(tracks_dir, analyze_dir, mann_coef) as follow:

cluster = SLURMCluster(cores=16,
                       processes=1,
                       memory="500GB",
                       account="nos-surge", # "nos-surge" for Hercules, "compute" for PW 
                       walltime="04:00:00",
                       interface="eth0", # only for PW
                       header_skip=['--mem'])
cluster.scale(6)
client = Client(cluster) 

And then running the script manually.

Could you please add this to the main script, and maybe with user defined inputs in input.conf? thanks

@SorooshMani-NOAA
Copy link
Collaborator

sure @FariborzDaneshvar-NOAA. One thing we need to test though, is what happens if we run this Dask code in the singularity container. I'm not sure how that would work with running new instances, etc.

@SorooshMani-NOAA SorooshMani-NOAA self-assigned this Apr 11, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants