You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi @voutcn,
is there a way to estimate the minimum mem requirement to process a dataset like this:
[read_lib_functions-inl.h : 209] Lib 0 (MetaGSB_022018_FD_input.corr.fastq.gz): interleaved, 216053580 reads, 150 max length
[read_lib_functions-inl.h : 209] Lib 1 (SB02metaG_FD_input.corr.fastq.gz): interleaved, 526835864 reads, 150 max length
[read_lib_functions-inl.h : 209] Lib 2 (SB12metaG_FD_input.corr.fastq.gz): interleaved, 362498508 reads, 150 max length
[read_lib_functions-inl.h : 209] Lib 3 ( SI3LmetaG_FD_input.corr.fastq.gz): interleaved, 376801276 reads, 150 max length
[read_lib_functions-inl.h : 209] Lib 4 (SI3UmetaG_FD_input.corr.fastq.gz): interleaved, 589974120 reads, 150 max length
I've been trying to get this running on a new compute cluster for a while now, and it always crashed/is terminated by the system because of memory usage. My previous system had 500gb RAM, the one now has nodes with 128gb each.
I've tried different things with --mem-flag 0 and -m 0.25 or --mem-flag 0 and -m 0.85 (and others), but obviously (and maybe not surprisingly) one node doesn't seem sufficient.
Thanks for your help!
The text was updated successfully, but these errors were encountered:
Megahit will read all reads into memory for graph building. Minimum memory required can be calculated by: N / 4 + n * 16 + M, where N is the total number of bases of the input reads, n is the number of reads and M is the working memory for sorting k-mers, which is related to the distribution of the memory but usually much less than N / 4 + n * 16.
Try -m 0.99 --mem-flag 0 and if it does not work, one node is insufficient.
Hi @voutcn,
is there a way to estimate the minimum mem requirement to process a dataset like this:
[read_lib_functions-inl.h : 209] Lib 0 (MetaGSB_022018_FD_input.corr.fastq.gz): interleaved, 216053580 reads, 150 max length
[read_lib_functions-inl.h : 209] Lib 1 (SB02metaG_FD_input.corr.fastq.gz): interleaved, 526835864 reads, 150 max length
[read_lib_functions-inl.h : 209] Lib 2 (SB12metaG_FD_input.corr.fastq.gz): interleaved, 362498508 reads, 150 max length
[read_lib_functions-inl.h : 209] Lib 3 ( SI3LmetaG_FD_input.corr.fastq.gz): interleaved, 376801276 reads, 150 max length
[read_lib_functions-inl.h : 209] Lib 4 (SI3UmetaG_FD_input.corr.fastq.gz): interleaved, 589974120 reads, 150 max length
I've been trying to get this running on a new compute cluster for a while now, and it always crashed/is terminated by the system because of memory usage. My previous system had 500gb RAM, the one now has nodes with 128gb each.
I've tried different things with --mem-flag 0 and -m 0.25 or --mem-flag 0 and -m 0.85 (and others), but obviously (and maybe not surprisingly) one node doesn't seem sufficient.
Thanks for your help!
The text was updated successfully, but these errors were encountered: