-
Notifications
You must be signed in to change notification settings - Fork 179
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
SLURM metagenome run - increased batMemory from restart not recognised #1281
Comments
Can't believe I missed it, but I just found issue #1253, so will try deleting 4-unitigger. |
Yes, that was going to be my suggestion, in general the scripts aren't re-generated if they exist so new parameters don't overwrite the old ones. |
Hello skoren, brilliant thanks - it is happily fat bogarting now:
It is great that it restarts so smoothly, providing great flexibility for the tricky samples. Sorry for the issue duplication! And thanks again! |
Sergey, I'd really appreciate your advice on memory. Unfortunately, for the really big metagenome (7000x coverage of 5m, 50 Gbp of data), even batMemory=498 wasn't enough... do you have any suggestions? Our grid nodes don't get bigger that 500GB, so I get the Otherwise, we have a single server with 1TB, so moving the data there might be the best option - though maybe this sample is just too big... Thank you! |
The issue is from this bit of logging:
where bogart thinks, based on a genome size of 5 Mbp, that you've really got 2810x coverage, and so it wants to load at least 'gobs and gobs' of overlaps per read. A simple fix would be to increase the genome size supplied to bogart (-gs 5000000). Increasing to 50 Mbp would reduce 'coverage' to 281x, and bogart would require only 552 overlaps per read. It picks the 'strongest' overlaps, and just ignores any additional overlaps. The genome size here is used only to compute coverage and N50 statistics in logs. And since we don't rewrite scripts, you can edit unitigger.sh and just restart canu. ;-) |
That's really awesome, thanks so much Brian :D |
Hello,
Firstly, thank you for a fantastic tool - it has worked very well so far, however I have run into some problems for a couple of our larger metagenomes.
I ran the Canu 1.8 command (see below) firstly calling our HPC slim nodes, with batMemory=200, and bogart failed due to memory limits.
When I increased the batMemory=450 and selected to use fat nodes instead, it failed at the same step and looking at the canu.out file (with the restart time stamp) it seem like the memory increase was not registered (still running with batMemory=200). It was trying to run the same command again. However, unitigger.jobSubmit-01.sh did seem to recognise the changed arguments... (see below)
This is my first time using an HPC cluster, so I am not sure if it is a problem with:
Here is the sbatch file I ran after changing the batMemory, and from slim to fat nodes:
Here is the error:
The coverage is very high... the other metagenome that failed is 7000x coverage for the setting of a 5m genome. The sample is ~50 Gbp of nanopore data.
The error from 4-unitigger/unitigger.err (the time stamp, similar to canu.out, indicates it was from the restart):
Config section of SLURM out
And output of unitigger.jobSubmit-01.sh:
And unitigger.out, just in case:
Thank you in advance for any suggestions,
Caitlin
Also, in case the restart canu and bogart logs (from canu-logs) are useful:
bogart log:
Canu log:
The text was updated successfully, but these errors were encountered: