New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Days spent on chomId_006_chr7_0013 #168
Comments
Hey, The last entry in the vcf from that region is
And you can see, that the DP value is already through the roof (for a WGS). I tried to supply a bed file with centromeres excluded
But it still takes MUCH longer in these segments in contrast to all other.
And all of those regions contain the centromere of their respective chromosome. I do suspect, that there is something, where the bailout, that surely happens for the other chromosomes, just doesnt happen here. It would be great if there was a fix for this. |
Leaving a comment to vote for a solution to this issue. We've run into this problem frequently though it seems somewhat arbitrary which samples take the longest to run, and it seems the bin that contains the centromere on Chr4 consistently takes the longest to run. |
I actually found a solution for me by including
into the strelka.ini Obviously you need to adjust the depth depending on your input. |
Thanks, I'll give that a try. |
Hello, and thank you for your tool!
When I run Strelka2 somatic with 24+ cores and 32G+ memory for 30x WGS data (with hg38 masked as recommended in the User Guide), ~1/2 of the samples get stuck on
Longest ongoing running task name: 'CallGenome+callGenomeSegment_chromId_006_chr7_0013'
for multiple hours, while the other jobs finish in a couple hours. >1/3 of the samples (out of ~70) keep running for >a week, until I shut them down. Restarting the job from where it left off does not fix the problem. Is this a bug you could please address? I can't share the data, but I don't find any regions of abnormally high depth on chr7.
Configuration example:
/path/to/configureStrelkaSomaticWorkflow.py
--normalBam /path/to/normal.bam
--tumorBam /path/to/tumor.bam
--referenceFasta /path/to/hg38.fa
--runDir /path/to/dir
--callRegions /path/to/strelka2-provided-call-regions-for-hg38.bed.gz
Run example (on HPC cluster node):
/path/to/runWorkflow.py -m local -j 24
Log example:
...[156412_1] [WorkflowRunner] [StatusUpdate] Workflow specification is complete?: True
...[156412_1] [WorkflowRunner] [StatusUpdate] Task status (waiting/queued/running/complete/error): 8/0/1/572/0
...[156412_1] [WorkflowRunner] [StatusUpdate] Longest ongoing queued task time (hrs): 0.0000
...[156412_1] [WorkflowRunner] [StatusUpdate] Longest ongoing queued task name: ''
...[156412_1] [WorkflowRunner] [StatusUpdate] Longest ongoing running task time (hrs): 17.6356
...[156412_1] [WorkflowRunner] [StatusUpdate] Longest ongoing running task name: 'CallGenome+callGenomeSegment_chromId_006_chr7_0013'
The text was updated successfully, but these errors were encountered: