Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Download and multiprocessing updates #226

Merged
merged 4 commits into from
May 4, 2017
Merged

Conversation

TimoRoth
Copy link
Member

@TimoRoth TimoRoth commented May 3, 2017

Goes back to the previous download behavior, while keeping the changed config intact.

While testing, I also found that it'd be nice to see which glacier crashed in which task while multiprocessing, so I added a wrapper-exception around it with the rgi_id and task name if those are available.

@fmaussion
Copy link
Member

While testing, I also found that it'd be nice to see which glacier crashed in which task while multiprocessing, so I added a wrapper-exception around it with the rgi_id and task name if those are available.

Yes, there are plenty of things that would be nice to know after run. I forgot to write them down but I'm going to do new runs soon and maybe we can define a strategy for better runs.

One of the things I'd really like to have is a list of successfully applied tasks on a glacier. Then, if you want to apply the task again on the same glacier it is ignored, unless forced by the user.

@fmaussion
Copy link
Member

I'm not sure why the test is suddenly failing, I think we can merge this

@TimoRoth
Copy link
Member Author

TimoRoth commented May 3, 2017

I have one more idea for an addition, the test failure seems unrelated.

@TimoRoth
Copy link
Member Author

TimoRoth commented May 3, 2017

This adds a parameter to override the cache layout, so if some data files are usually stored in a specific format, it can now be defined without having to manually intervene or copy the files around.

@fmaussion
Copy link
Member

@TimoRoth if you have time could you look at this PR on xarray? pydata/xarray#1393

Here they implement a relatively simple checksum to make the algorithm more robust. Like Stephan I am quite surprised as to why it's not implemented in urlretrieve per default, but if it can make our downloads more robust...

@TimoRoth
Copy link
Member Author

TimoRoth commented May 3, 2017

That's a special case for downloads from github, as github can give you a checksum for every file.
As most (all but one) of our downloads are not from github, we can't do something like that.

@fmaussion fmaussion merged commit c7a08b9 into OGGM:master May 4, 2017
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants