You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I'm also having the issue of not enough disk space. I'm working on a plant genome of ~600M with 7X coverage (more data is being produced), and 7 Tb /scratch disk space was used up.
I have read issue #587 and #703 and I saw the Sergey's answers. I wonder if it was possible to estimate disk usage. Now I'm running serveral tests with different number of input reads and monitorting disk usage, but I want to ask the question here to see if there is a better way.
Thank you very much,
Zhu
The text was updated successfully, but these errors were encountered:
7x really isn't enough to do anything useful, and could end up using more space than with higher coverage since the repeats won't be masked out as heavily. It would be better to estimate based on the output of a few jobs from the full read set, then adjusting parameters if needed.
Space usage depends heavily on genome properties (repeats, duplications, ploidy) and read length/quality. Plants, unfortunately, seem to want a lot of space. 10TB is a reasonable guess. I've seen it as bad as 20TB, but I forget what the genome was.
Hello,
I'm also having the issue of not enough disk space. I'm working on a plant genome of ~600M with 7X coverage (more data is being produced), and 7 Tb /scratch disk space was used up.
I have read issue #587 and #703 and I saw the Sergey's answers. I wonder if it was possible to estimate disk usage. Now I'm running serveral tests with different number of input reads and monitorting disk usage, but I want to ask the question here to see if there is a better way.
Thank you very much,
Zhu
The text was updated successfully, but these errors were encountered: