Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Running on a SLURM cluster, gives a lot of errors. #32

Closed
sameet opened this issue Sep 14, 2017 · 45 comments
Closed

Running on a SLURM cluster, gives a lot of errors. #32

sameet opened this issue Sep 14, 2017 · 45 comments

Comments

@sameet
Copy link

sameet commented Sep 14, 2017

Hi,

I have the following directory structure:

references:
total 8374528
-rwxr-xr-x+ 1 sm2556 mane 3157608038 Sep 13 12:32 Homo_sapiens_assembly19.fasta  
-rw-r--r--+ 1 sm2556 mane       6663 Sep 13 13:31 Homo_sapiens_assembly19.fasta.amb
-rw-r--r--+ 1 sm2556 mane        939 Sep 13 13:31 Homo_sapiens_assembly19.fasta.ann
-rw-r--r--+ 1 sm2556 mane 3095694072 Sep 13 13:30 Homo_sapiens_assembly19.fasta.bwt
-rw-r--r--+ 1 sm2556 mane  773923497 Sep 13 13:31 Homo_sapiens_assembly19.fasta.pac
-rw-r--r--+ 1 sm2556 mane 1547847040 Sep 13 13:44 Homo_sapiens_assembly19.fasta.sa
-rw-r--r--+ 1 sm2556 mane        377 Sep 13 15:19 Homo_sapiens_assembly19.sizes

restriction_sites:
total 15360
-rw-r--r--+ 1 sm2556 mane 7762896 Sep 13 11:45 hg19_HindIII_new.txt
-rw-r--r--+ 1 sm2556 mane 7762896 Sep 13 11:45 hg19_HindIII.txt

scripts:
total 92800
-rwxr-xr-x+ 1 sm2556 mane     3519 Sep 13 11:26 check.sh
-rwxr-xr-x+ 1 sm2556 mane    15349 Sep 13 11:26 chimeric_blacklist.awk
-rwxr-xr-x+ 1 sm2556 mane     1971 Sep 13 11:26 cleanup.sh
-rwxr-xr-x+ 1 sm2556 mane     3584 Sep 13 11:26 collisions.awk
-rwxr-xr-x+ 1 sm2556 mane     1616 Sep 13 11:26 countligations.sh
-rwxr-xr-x+ 1 sm2556 mane    13448 Sep 13 11:26 diploid.pl
-rw-r--r--+ 1 sm2556 mane     2449 Sep 13 11:26 diploid_split.awk
-rwxr-xr-x+ 1 sm2556 mane     5325 Sep 13 11:26 dups.awk
-rw-r--r--+ 1 sm2556 mane     3726 Sep 13 11:26 fragment_4dnpairs.pl
-rwxr-xr-x+ 1 sm2556 mane     3711 Sep 13 11:26 fragment.pl
-rw-r--r--+ 1 sm2556 mane 30745856 Sep 13 12:31 juicebox
-rw-r--r--+ 1 sm2556 mane 30745856 Sep 13 12:30 Juicebox.jar
-rw-r--r--+ 1 sm2556 mane 30751431 Sep 13 12:30 juicebox_tools.7.0.jar
-rwxr-xr-x+ 1 sm2556 mane     2388 Sep 13 11:26 juicer_arrowhead.sh
-rwxr-xr-x+ 1 sm2556 mane     3269 Sep 13 11:26 juicer_hiccups.sh
-rwxr-xr-x+ 1 sm2556 mane     3651 Sep 13 11:26 juicer_postprocessing.sh
-rwxr-xr-x+ 1 sm2556 mane    41529 Sep 13 11:26 juicer.sh
-rwxr-xr-x+ 1 sm2556 mane     4659 Sep 13 11:26 LibraryComplexity.class
-rwxr-xr-x+ 1 sm2556 mane     7204 Sep 13 11:26 LibraryComplexity.java
-rwxr-xr-x+ 1 sm2556 mane     2354 Sep 13 11:26 makemega_addstats.awk
-rwxr-xr-x+ 1 sm2556 mane    12782 Sep 13 11:26 mega.sh
-rwxr-xr-x+ 1 sm2556 mane     2455 Sep 13 11:26 relaunch_prep.sh
-rwxr-xr-x+ 1 sm2556 mane     5200 Sep 13 11:26 split_rmdups.awk
-rwxr-xr-x+ 1 sm2556 mane    14572 Sep 13 11:26 statistics.pl
-rwxr-xr-x+ 1 sm2556 mane     1751 Sep 13 11:26 stats_sub.awk

fastq:
total 0
lrwxrwxrwx 1 sm2556 mane 67 Sep 13 15:33 S1_003_HiC_R1.fastq.gz -> ../../analysis jul052016/S1_003_HiC/Unaligned/S1_003_HiC_1.fastq.gz
lrwxrwxrwx 1 sm2556 mane 67 Sep 13 15:34 S1_003_HiC_R2.fastq.gz -> ../../analysis-jul052016/S1_003_HiC/Unaligned/S1_003_HiC_2.fastq.gz

My run.sh script for the SLURM batch submission looks as follows:

#!/bin/bash
#SBATCH --partition=general
#SBATCH --job-name=Juicer
#SBATCH --ntasks=1 --nodes=1
#SBATCH --mem-per-cpu=6000
module load BWA; module load Java;  bash /home/sm2556/project/hic-golden-uconn-feb0222016/hic-analysis-sept142017/scripts/juicer.sh -g hg19 -d /home/sm2556/project/hic-golden-uconn-feb0222016/hic-analysis-sept142017 -q general -l general -a 'Reference' -S 'early' -p /home/sm2556/project/hic-golden-uconn-feb022216/hic-analysis-sept142017/references/Homo_sapiens_assembly19.sizes -s 'HindIII' -y /home/sm2556/project/hic-golden-uconn-feb0222016/hic-analysis-sept142017/restriction_sites/hg19_HindIII.txt - D /home/sm2556/project/hic-golden-uconn-feb0222016/hic-analysis-sept142017 -x

The scripts folder was copied from the cloned GitHub repository of the juicer/SLURM/scripts.

I get tons of error messages about dependencies not being satisfied, but I still get the part of script that "split" the fastq.gz file correctly, but still ends with error. The actual bwa mem call never happens on the cluster. When I tried to run the script in the CPU mode it started the alignment. But my files are too big, and CPU mode will take a long time. Am I doing something wrong?

@nchernia
Copy link
Collaborator

nchernia commented Sep 14, 2017 via email

@sameet
Copy link
Author

sameet commented Sep 14, 2017

Hi @nchernia,
Thank you for the prompt reply. I fixed the run.sh as you suggested. Also added the edited the load_bwa, and load_java in the juicer.sh. I removed the debug, aligned , and splits folders (want to start with a clean slate). I re-ran the run.sh as follows:

[sm2556@ruddle2 hic-analysis-sept142017]$ sbatch run.sh

[sm2556@ruddle2 hic-analysis-sept142017]$ more slurm-1674096.out
(-: Looking for fastq files...fastq files exist
(-: Aligning files matching /home/sm2556/project/hic-golden-uconn-feb0222016/hic-analysis-sept142017/fastq/*_R*.fastq* 
in queue general to genome hg19 with no fragment delimited maps.
(-: Created /home/sm2556/project/hic-golden-uconn-feb0222016/hic-analysis-sept142017/splits and /home/sm2556/project/hic-golden-uconn-feb0222016/hic-analysis-sept142017/aligned.
(-: Starting job to launch other jobs once splitting is complete
sbatch: error: Batch job submission failed: Requested node configuration is not available
sbatch: error: Batch job submission failed: Requested node configuration is not available
sbatch: error: Batch job submission failed: Job dependency problem
sbatch: error: Batch job submission failed: Job dependency problem
sbatch: error: Batch job submission failed: Job dependency problem
sbatch: error: Batch job submission failed: Job dependency problem
sbatch: error: Batch job submission failed: Job dependency problem
sbatch: error: Batch job submission failed: Job dependency problem
sbatch: option requires an argument -- 'd'
(-: Finished adding all jobs... Now is a good time to get that cup of coffee..

I can see the splits directory re-created, and bunch of .fastq chunks being created in it right now. But I am pretty sure that the run is going to fail. How do I fix this?

@nchernia
Copy link
Collaborator

nchernia commented Sep 14, 2017 via email

@sameet
Copy link
Author

sameet commented Sep 14, 2017

#!/bin/bash
#SBATCH --partition=general
#SBATCH --job-name=Juicer
#SBATCH --ntasks=1 --nodes=1
#SBATCH --mem-per-cpu=6000

bash /home/sm2556/project/hic-golden-uconn-feb0222016/hic-analysis-sept142017/scripts/juicer.sh -g hg19 -d /home/sm2556/project/hic-golden-uconn-feb0222016/hic-analysis-sept142017 -q general -l general -a 'Reference' -S 'early' -p /home/sm2556/project/hic-golden-uconn-feb022216/hic-analysis-sept142017/references/Homo_sapiens_assembly19.sizes -s 'HindIII' -y /home/sm2556/project/hic-golden-uconn-feb0222016/hic-analysis-sept142017/restriction_sites/hg19_HindIII.txt -D /home/sm2556/project/hic-golden-uconn-feb0222016/hic-analysis-sept142017 -x

That is my run.sh.

@nchernia
Copy link
Collaborator

nchernia commented Sep 14, 2017 via email

@sameet
Copy link
Author

sameet commented Sep 14, 2017

Hi,

Before I answer your question, I think I found one mistake that I was doing. When I call the script without giving number of threads, it assumes 16, and allocates 40 GB per core. We do not have many such nodes. So that seems to be one of the problems. I have edited the script to use -t 5, and hopefully that will solve one problem.

Also I, with help from our system administrator, went through your source code carefully, and I finally understood why I do not need to call my run.sh by sbatch. The original script of course died after splitting the fastq files into 128 chunks. Now in my splits directory I have 128 x 2 fastq files. At this point I just re-ran the actual script part of run.sh, and it is generating lots of .txt files in the splits folder, like so:

-rw-r--r--+ 1 sm2556 mane 9 Sep 14 15:53 S1_003_HiC.fastq102.fastq_linecount.txt
-rw-r--r--+ 1 sm2556 mane 8 Sep 14 15:53 S1_003_HiC.fastq102.fastq_norm.txt.res.txt
-rw-r--r--+ 1 sm2556 mane 9 Sep 14 15:53 S1_003_HiC.fastq103.fastq_linecount.txt
-rw-r--r--+ 1 sm2556 mane 8 Sep 14 15:53 S1_003_HiC.fastq103.fastq_norm.txt.res.txt
-rw-r--r--+ 1 sm2556 mane 9 Sep 14 15:53 S1_003_HiC.fastq104.fastq_linecount.txt
-rw-r--r--+ 1 sm2556 mane 8 Sep 14 15:53 S1_003_HiC.fastq104.fastq_norm.txt.res.txt

And, my debug folder has a bunch of .err, and .out files.

In the debug folder, there is a head-1674970.out file, the contents are as follows:

Thu Sep 14 16:11:33 EDT 2017
Experiment description: Reference; Juicer version 1.5.6;5 threads; splitsize 90000000; openjdk 
version "1.8.0_131"; ./scripts/juicer.sh -t 5 -g hg19 -d /home/sm2556/project/hic-golden-uconn-feb0222016/hic-analysis-sept142017 -q general -l general -a Reference -S early -p ./references/Homo_sapiens_assembly19.sizes -s HindIII -y ./restriction_sites/hg19_HindIII.txt -D /home/sm2556/project/hic-golden-uconn-feb0222016/hic-analysis-sept142017 -x

@sameet
Copy link
Author

sameet commented Sep 14, 2017

Is there any place where I can see the batch file generated with this. It seems that the first step where it splits the fastq into multiple chunks, seems to be working, but not the alignment part. What am I missing? When I try to do this in CPU mode, that seems to work.

@nchernia
Copy link
Collaborator

nchernia commented Sep 14, 2017 via email

@sameet
Copy link
Author

sameet commented Sep 15, 2017

==> count_ligation-1676612.out <==
Thu Sep 14 18:31:34 EDT 2017
Thu Sep 14 18:36:55 EDT 2017

==> head-1676609.out <==
Thu Sep 14 18:31:34 EDT 2017
Experiment description: Reference; Juicer version 1.5.6;5 threads; splitsize 90000000; openjdk 
version "1.8.0_131"; ./scripts/juicer.sh -d /ycga-gpfs/project/ycga/mane/sm2556/hic-golden-
uconn-feb0222016/hic-analysis-sept152017 -q general -l general -s HindIII -a Reference -p 
./references/genome.chrom.sizes -y ./restriction_sites/hg19_HindIII.txt -D /ycga-
gpfs/project/ycga/mane/sm2556/hic-golden-uconn-feb0222016/hic-analysis-sept152017 -z 
./references/genome.fa -t 5

==> split-1676610.out <==
Split file: S1_003_HiC_R1.fastq
Thu Sep 14 20:28:23 EDT 2017

==> split-1676611.out <==
Split file: S1_003_HiC_R2.fastq
Thu Sep 14 20:25:00 EDT 2017

Today I saw these (this is just an example I have ~200 such files in the debug folder.

==> count_ligation-1676612.out <==
Thu Sep 14 18:31:34 EDT 2017
Thu Sep 14 18:36:55 EDT 2017

==> count_ligation-1677151.out <==
Fri Sep 15 08:19:21 EDT 2017
Fri Sep 15 08:24:16 EDT 2017

==> count_ligation-1677155.out <==
Fri Sep 15 08:19:22 EDT 2017
Fri Sep 15 08:24:13 EDT 2017

==> count_ligation-1677159.out <==
Fri Sep 15 08:19:22 EDT 2017
Fri Sep 15 08:24:19 EDT 2017

==> count_ligation-1677163.out <==
Fri Sep 15 08:19:22 EDT 2017
Fri Sep 15 08:24:19 EDT 2017

==> count_ligation-1677167.out <==
Fri Sep 15 08:19:22 EDT 2017
Fri Sep 15 08:24:19 EDT 2017

Me and our system admins were at this issue for nearly whole of yesterday, but to the best of my understanding, the juicer.sh should have found everything required.

@nchernia
Copy link
Collaborator

nchernia commented Sep 15, 2017 via email

@sameet
Copy link
Author

sameet commented Sep 15, 2017

No, there are some 200 odd files. The output in all of them looks similar. There are two other types, head.*, and split-*. Those look as follows:

==> head-1676609.out <==
Thu Sep 14 18:31:34 EDT 2017
Experiment description: Reference; Juicer version 1.5.6;5 threads; splitsize 90000000; openjdk version "1.8.0_131"; ./scripts/juicer.sh -d /ycga-gpfs/project/ycga/mane/sm2556/hic-golden-uconn-feb0222016/hic-analysis-sept152017 -q general -l general -s HindIII -a Reference -p ./references/genome.chrom.sizes -y ./restriction_sites/hg19_HindIII.txt -D /ycga-gpfs/project/ycga/mane/sm2556/hic-golden-uconn-feb0222016/hic-analysis-sept152017 -z ./references/genome.fa -t 5

==> head-1677150.out <==
Fri Sep 15 08:19:21 EDT 2017
Experiment description: Reference; Juicer version 1.5.6;5 threads; splitsize 90000000; openjdk version "1.8.0_131"; ./scripts/juicer.sh -d /ycga-gpfs/project/ycga/mane/sm2556/hic-golden-uconn-feb0222016/hic-analysis-sept152017 -q general -l general -s HindIII -a Reference -p ./references/genome.chrom.sizes -y ./restriction_sites/hg19_HindIII.txt -D /ycga-gpfs/project/ycga/mane/sm2556/hic-golden-uconn-feb0222016/hic-analysis-sept152017 -z ./references/genome.fa -t 5

==> head-1677810.out <==
Fri Sep 15 08:25:09 EDT 2017
Experiment description: Reference; Juicer version 1.5.6;5 threads; splitsize 90000000; openjdk version "1.8.0_131"; ./scripts/juicer.sh -d /ycga-gpfs/project/ycga/mane/sm2556/hic-golden-uconn-feb0222016/hic-analysis-sept152017 -q general -l general -s HindIII -a Reference -p ./references/genome.chrom.sizes -y ./restriction_sites/hg19_HindIII.txt -D /ycga-gpfs/project/ycga/mane/sm2556/hic-golden-uconn-feb0222016/hic-analysis-sept152017 -z ./references/genome.fa -t 5

==> split-1676610.out <==
Split file: S1_003_HiC_R1.fastq
Thu Sep 14 20:28:23 EDT 2017

==> split-1676611.out <==
Split file: S1_003_HiC_R2.fastq
Thu Sep 14 20:25:00 EDT 2017

@nchernia
Copy link
Collaborator

nchernia commented Sep 15, 2017 via email

@sameet
Copy link
Author

sameet commented Sep 15, 2017

Nope.

@nchernia
Copy link
Collaborator

nchernia commented Sep 15, 2017 via email

@sameet
Copy link
Author

sameet commented Sep 15, 2017

log.txt
I did the sh juicer.sh ... >log 2&>1 this time. The errors look the same. The jobs got fired on the cluster and also died within few minutes. I am pasting the log file here.

@nchernia
Copy link
Collaborator

nchernia commented Sep 15, 2017 via email

@sameet
Copy link
Author

sameet commented Sep 15, 2017

Understood. I will try that immediately. All of our nodes have at least 128 GB RAM. We have few nodes with upto 1.5 TB RAM. But setting t=1 makes sense. I will try that and let you know how that works.

@sameet
Copy link
Author

sameet commented Sep 15, 2017

Well, something more seems to have happened 😄 . The bwa mem jobs at least seem to have been generated and submitted. But the outputs are still empty. I am attaching the tail -n 5 align*.out herewith.
align_log.txt

@nchernia
Copy link
Collaborator

nchernia commented Sep 15, 2017 via email

@sameet
Copy link
Author

sameet commented Sep 15, 2017

There are bunch of .sam files. But they are empty.

@sameet
Copy link
Author

sameet commented Sep 15, 2017

This is the last few lines of ls -l in the splits directory.

-rw-r--r--+ 1 sm2556 mane 5.4G Sep 14 20:24 S1_003_HiC_R2.fastq127.fastq
-rw-r--r--+ 1 sm2556 mane    0 Sep 15 09:51 S1_003_HiC_R2.fastq127.fastq.sam
-rw-r--r--+ 1 sm2556 mane    0 Sep 15 09:52 S1_003_HiC_R2.fastq127.fastq_sort1.sam
-rw-r--r--+ 1 sm2556 mane    0 Sep 15 09:52 S1_003_HiC_R2.fastq127.fastq_sort.sam
-rw-r--r--+ 1 sm2556 mane 741M Sep 14 20:25 S1_003_HiC_R2.fastq128.fastq
-rw-r--r--+ 1 sm2556 mane    0 Sep 15 09:52 S1_003_HiC_R2.fastq128.fastq.sam
-rw-r--r--+ 1 sm2556 mane    0 Sep 15 09:52 S1_003_HiC_R2.fastq128.fastq_sort1.sam
-rw-r--r--+ 1 sm2556 mane    0 Sep 15 09:52 S1_003_HiC_R2.fastq128.fastq_sort.sam

@nchernia
Copy link
Collaborator

nchernia commented Sep 15, 2017 via email

@nchernia
Copy link
Collaborator

nchernia commented Sep 15, 2017 via email

@sameet
Copy link
Author

sameet commented Sep 15, 2017

The align*.err showed the following:

==> align2-1679007.err <==
slurmstepd: error: execve(): bwa: No such file or directory
srun: error: c17n06: task 0: Exited with exit code 2

So I checked the following:

[******@****** hic-analysis-sept152017]$ grep load_bwa ./scripts/juicer.sh
    # load_bwa="module load BioBuilds/2015.04"
    load_bwa="module load BWA"
        $load_bwa
                $load_bwa
                $load_bwa
[sm2556@ruddle2 hic-analysis-sept152017]$ module load BWA
[******@****** hic-analysis-sept152017]$ which bwa
/ycga-gpfs/apps/hpc/software/BWA/0.7.15-foss-2016a/bin/bwa
[******@****** hic-analysis-sept152017]$

@nchernia
Copy link
Collaborator

nchernia commented Sep 15, 2017 via email

@sameet
Copy link
Author

sameet commented Sep 15, 2017

Yes, I changed it. As you can see in the second panel, the commented out line is the original line from the juicer.sh. The second line is the one that I put in.

@nchernia
Copy link
Collaborator

nchernia commented Sep 15, 2017 via email

@nchernia
Copy link
Collaborator

nchernia commented Sep 15, 2017 via email

@sameet
Copy link
Author

sameet commented Sep 15, 2017

I figured out one reason, inside the script it is using $load_bwa, that was failing, I changed it to ${load_bwa}, that seems to be working, but I got bus and memory errors. I will work with my system admin to see if anything can be done about that.

@nchernia
Copy link
Collaborator

nchernia commented Sep 15, 2017 via email

@sameet
Copy link
Author

sameet commented Sep 15, 2017

Thanks. Also setting -t 2 was required because of memory issues. I guess general rule of thumb is the memory allocated should be more than the size of the chunk. It is aligning now. I will keep you posted in this thread on how things go.

@sameet
Copy link
Author

sameet commented Sep 17, 2017

The things seem to be working so far. The alignments completed. I see a lot of merge* files in the debug folder. The run ended without completion or any files in the aligned folder. After reading your wiki, I thought this would be the time when I start re-launch on the juicer.sh with -S merge option. I changed the partition parameters to allocate nodes with maximum RAM and the jobs seem to be working. I can see merged_sort.txt in the aligned folder. I guess this will be a fairly slow step. Shall keep you posted on how things go.

I noticed in the code that the time limit for the merge step is 24 hours. Isn't that too low? I think that most of the times 7 days is a good limit for batch jobs.

@sameet
Copy link
Author

sameet commented Oct 2, 2017

After one more false start, the pipeline seems to be working. I am currently at the dedup step. The file was huge. It was split into over 500 chunks. One of the chunks failed. Will that affect the final step. Will I have to run dedup again? Can I just remove that one .err file, and let it continue?

@nchernia
Copy link
Collaborator

nchernia commented Oct 2, 2017 via email

@sameet
Copy link
Author

sameet commented Oct 2, 2017

That may not be an option, because the error message was it ran out of time on the node, took more than 24 hours. I fear that if i run it again, the same is going to happen. Is there a work around. Following is the error message for that chunk in the debug directory:

slurmstepd: error: *** JOB 1713635 ON bigmem01 CANCELLED AT 2017-09-30T06:39:21 DUE TO TIME LIMIT ***

Is there a way around this?

@nchernia
Copy link
Collaborator

nchernia commented Oct 2, 2017 via email

@sameet
Copy link
Author

sameet commented Oct 2, 2017

Hi,
there is no attachment.

Sameet

@nchernia
Copy link
Collaborator

nchernia commented Oct 2, 2017 via email

@sameet
Copy link
Author

sameet commented Oct 26, 2017

Most of the pipeline ran till the deduplication step. I got the following error in the hic step

 more hic-1692493.err
Problem with creating fragment-delimited maps, NullPointerException. 
This could be due to a null fragment map or to a mismatch in the chromosome name in the 
 fragment map v is-a-vis the input file or chrom.sizes file. 
Exiting.

Is there any way to fix this.

@nchernia
Copy link
Collaborator

nchernia commented Oct 27, 2017 via email

@sameet
Copy link
Author

sameet commented Oct 27, 2017

I think the error was generated before the .hic files were generated. Everystep after dedup has failed. The dedup worked fine.

@nchernia
Copy link
Collaborator

nchernia commented Oct 27, 2017 via email

@Aannaw
Copy link

Aannaw commented Jan 18, 2022

I think the error was generated before the .hic files were generated. Everystep after dedup has failed. The dedup worked fine.
Hello
have you solved the problem and successfully run juicer with SLURM. I also want to run juicer with SLURM but I am not root user, and I can not apply module load to change $load_bwa, $load_java,$load_gnu. I am not sure if I can set the path of software like load_bwa="export PATH=/data/01/user157/software/bwa:$PATH" , load_java="export PATH=/usr/bin/:$PATH" and load gpu=load_java=" export PATH=/usr/bin/:$PATH" .
I would appreciate with it if you could give me any suggestions!
Best wishes!

@sa501428
Copy link
Member

If you are using a recent jar, these issues are all resolved. You should contact the SLURM cluster manager and ask how to load bwa/java/etc. Those commands can then be replaced to be specific to your instance. If you have further questions, see the forum.

@Aannaw
Copy link

Aannaw commented Jan 18, 2022

@sa501428
Hello Professor
In the end I chose to link the CPU directory to my work directory and then run with CPU. The command is "./scripts/juicer.sh -g Ma6.genome -d /data/01/user157/HIFI/Hic-anchor/Ma6/juicer/work/hic_data -s MboI -p ./restriction_sites/Ma6.genome.chrom.sizes -y ./restriction_sites/Ma6.genome_MboI.txt -z ./references/Ma6.bp.p_ctg.fasta -D /data/01/user157/HIFI/Hic-anchor/Ma6/juicer/work -t 80 -e early". However, it seems that thehic raw readshave not been split in the tmp split directory. Also I can not find the debug directory. The following picture is the content of split directory. I just check the process and It is running with chimera read handling. I am really confused what is chimera read handling after looking through chimeric_sam.awk . Also, I refer to the juicer.sh but I can not find the chunk size to split like slurm.
I would appreciate it if you could give me any suggestions?
Best wishes.
juicer-split

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants