Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[FR] Allow resuming in fasterq-dump #841

Open
ArchKudo opened this issue Aug 17, 2023 · 3 comments
Open

[FR] Allow resuming in fasterq-dump #841

ArchKudo opened this issue Aug 17, 2023 · 3 comments

Comments

@ArchKudo
Copy link

I'm using fasterq-dump to download and extract fastq files of all runs (~250 files, 50GB each) of project PRJEB31266
However I find it very flaky, failing multiple times. Retrying it cause to download & extract the entire file from scratch which is wasteful in terms of bandwidth and processing.
It would be nice if it had something similar to --resume flag present in prefetch

@klymenko
Copy link
Contributor

Did you run prefetch before fasterq-dump?

@ArchKudo
Copy link
Author

Hello, thank you for replying.
I ended up using prefetch before fasterq-dump.
However, this didn't fix the problem since fasterq-dump was utilizing more space than what was available in tmp/ and required utilizing the --temp flag.

I still find value in implementing a resume feature, as the extraction process takes longer compared to downloading the sra file. Moreover, it might fail midway due to reasons such as running out of tmp/ storage. However, I leave it up to you to decide whether to close this issue.

@kcmtest
Copy link

kcmtest commented Oct 9, 2023

Hello, thank you for replying. I ended up using prefetch before fasterq-dump. However, this didn't fix the problem since fasterq-dump was utilizing more space than what was available in tmp/ and required utilizing the --temp flag.

I still find value in implementing a resume feature, as the extraction process takes longer compared to downloading the sra file. Moreover, it might fail midway due to reasons such as running out of tmp/ storage. However, I leave it up to you to decide whether to close this issue.

how did you fix im also similar situation where i have like more than 1000 files to download but it fails after 300-350 files, even i use prefetch first so in my case the prefetch is not able to get the given ID. Is it a space issue ? " it might fail midway due to reasons such as running out of tmp/ storage." can you explain

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants