-
Notifications
You must be signed in to change notification settings - Fork 502
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
EXITING because of FATAL ERROR in reads input: short read sequence line: 0 #493
Comments
Hi @juulluu21 this indicates some problem with formatting of the fastq files. Cheers |
Hi Alex, Thanks very much. I messed up during the adapter clipping step I guess. Now, its working fine. I have also changed the zcat to gunzip -c option. Here is my command: star --runThreadN 36 --genomeDir ./Genome --readFilesIn R1.fastq.gz. R2.fastq.gz --outFileNamePrefix R.sam --readFilesCommand gunzip -c I have two more questions.
Thanks a ton again, Alex! |
Hi @juulluu21 compared to the normal mapping rate >90% to a genome, you are getting low mappability. Cheers |
Actually most of the reads are coming from rRNA. My library prep didn't ribo depleted my sample. However, I see several mismatches in the reads (when I see them in IGV). Would you suggest some stringent mapping? I am also not very happy with the transcriptome that I am mapping against (I see many mismatches/SNPs when we sanger sequence some genes). Thanks very much Alex! |
Hi @juulluu21 ribosomal reads should appear as multimappers - if their sequences are present in the genome assembly. Cheers |
HI Alex, Thanks very much for your support! I am now adding using these parameters: My mapping has increased by 10%. But, still there are ~65% unmapped reads (% of reads unmapped: too short) If I use these parameters If I use these parameters
If I use these parameters Number of input reads | 29216568 **I am mapping the reads against transcriptome (small ~25000 assembled contigs). I do not have a genome. With outFilterScoreMinOverLread 0.1 and outFilterScoreMinOverLread 0.0, I guess I am getting most of the non-specific mapping. So, I guess I have to be satiefied with --outFilterScoreMinOverLread 0.3 --outFilterMatchNminOverLread 0.3 --outFilterMismatchNmax 20. Do you suggest to change between --outFilterScoreMinOverLread 0.3 --outFilterMatchNminOverLread 0.3 ?? Thanks a ton, Alex!** |
Hi @juulluu21 as you reduce the mapping stringency (mapped length) you can map more reads, however, only small portions of these reads are mapped, which cannot guarantee a correct mapping location. Effectively, you are trading off precision for higher sensitivity. You would have to think if your downstream analysis tolerates low precision (i.e. high false positive rate). In general, I would not recommend going below 0.6 for these parameters. The best solution to low mappability is to figure out why it happens. Cheers |
Thanks Alex. I will be satisfied with my ~30% mapping. Thanks!! |
Hello there, |
Hi @zillurbmb51 it looks like the fastqs are not properly formatted. Cheers |
These lines look OK. |
Hey Alex, Just had a follow-up question that was related to this post. I am having a similar error. Paired-end reads that have the same format as zillurbmb51. I have not tried your suggestion about trying a smaller set but I tried running them as single instead of paired reads and they aligned perfectly fine without errors. I was curious if you had seen this before? I also had issues with having multiple threads. my ulimit is unlimited but I was only able to run mapping jobs with 20 threads. Anything more and I would get a permissions error in creating files. For reference, I am running STAR 2.7.0f. Also just wanted to say that you do an amazing job providing support to us all! |
Hi @zehemminger the OP error did not get resolved, I cannot reproduce with my data. If you can find a small subset that still causes this error, please send it to me along with the Log.out file. Cheers |
I am getting a similar error. When running on fastq.gz files (I've confirmed are formatted correctly, I get):
I am running the command
And first few lines of a sample file:
|
Okay so when I run with
or
now I'm getting:
|
Hi Buuntu, please send me the Log.out file. Cheers |
So I was able to fix the error by using the option Thanks Alex |
Right - at the moment this parameter needs to be scaled down for small genomes. (See Manual 2.2.5). |
**Hi,
I know this is a very old question. Has it been solved yet?
I am running STAR on paired end reads:**
_star --runThreadN 50 --genomeDir index --readFilesIn ASS_1_1.fq.gz ASS__1_2.fq.gz --outFileNamePrefix ASS.mapped.sam --readFilesCommand gzcat
Sep 21 16:20:32 ..... started STAR run
Sep 21 16:20:33 ..... loading genome
Sep 21 16:20:39 ..... started mapping
EXITING because of FATAL ERROR in reads input: short read sequence line: 0
Read Name=@NB501259:103:HNTJLBGX2:1:21310:2440:4630
Read Sequence====
DEF_readNameLengthMax=50000
DEF_readSeqLengthMax=650
Sep 21 16:21:53 ...... FATAL ERROR, exiting
Segmentation fault: 11_
**I have cut my reads by cutadapt. Is there any option in STAR to get this problem around? Or, I have to do something with the cutadapt now?
Thanks.**
The text was updated successfully, but these errors were encountered: