Skip to content

Scripts to Analyze DTC Sequencing and Genotyping Data (and some comparisons to Veritas WGS data)

Notifications You must be signed in to change notification settings

cwarden45/DTC_Scripts

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

For Research Purposes Only!

Note: After attending the COH Clinical Cancer Genomics Conference, I discovered that I should be saying "Patient Initiated Testing" instead of "Direct To Consumer" (DTC, in the repository name) because there are some results that are signed off by an on-line physician (although I never talked to a medical professional before ordering/receiving any results).


The code sections of this repository mostly focus on analysis that you can perform on your local computer. While learning some coding to be able to extract additional information from your raw genomics data is a useful (and marketable) life skill, the comparison part (emphasized more in the subfolders) can at least partially be carried out through precisionFDA (possibly with some file formatting changes, like I did for my vcf files).

1) Creating a precisionFDA account is free and you can do it with your personal G-mail address (at least if you have genomic data to upload, like myself). The analysis run within precisionFDA is also free for users.

2) While there are cost and time limits to the total number of companies that I can check as an individual (for example, notice the time it took me to put together a response in this issue thread, admittedly needing to do a noticable portion of that outside working hours), I think there could be benefits for people to obtain raw data from at least twice (although I would also recommend free genotyping from Genes for Good; and I believe, if selected, some people can get free sequencing from All of US). This can give you a feel for the variety of results that you can receive, and I think precisionFDA provides a way that you can get a feeling of fulfillment for contributing to on-going surveillance of genomics companies (kind of like a donation to better society, with an understanding that accuracy can sometimes be harder to define than you might expect).

3) Of course, this requires companies to provide raw data. Likewise, I think it is worth noting that I could improve concordance between my Veritas WGS data and my Genos Exome data if I reprocessed the raw data (once you create an account, you can access my data and precisionFDA comparisons here). While I also applaud free interfaces to view data from companies (like data.color.com), I think this is a good example of why not providing raw data (I think FASTQ/BAM/VCF should be provided for high-throughput sequencing data) should be considered unacceptable.

Also, I believe certain things like using IGV to visualize your alignments (which is very important) should be done on your local computer (instead of on the cloud in precisionFDA).


I'm not exactly sure what can be currently downloaded from the providers, but you could potentially test the scripts with my data in the Personal Genome Project hu832966. I think the formatting is the same as I originally downloaded...


These are some scripts that I used to analyze my Veritas WGS (Whole Genome Sequencing) Data.

Step #1) Download .bam.tgz and .vcf.gz files

1a) Extract the per-chromosome .bam files (the [ID].bam.tgz file) into a folder called "chr_bam". On Windows, you can do this using 7zip.

Step #2) Download IGV: https://www.broadinstitute.org/software/igv/download

**Make sure you have JRE installed on your computer

At this point, you can already visualize your genomic alignment (separately for each chromosome)

Step #3) Download Docker and Dependencies from Docker Image: docker pull cwarden45/dnaseq-dependencies. For my version of Windows (Windows 10, Home Edition), I needed to install Docker Toolbox.

On Windows, if you run git clone https://github.com/cwarden45/DNAseq_templates in your Documents folder, then you will re-create the path I use in the next step (with your own username). Strictly speaking, you don't need git to run the next steps, but you will have to copy and paste the contents of each script you want to run (in a plain-text editor, like Notepad++). If you try to save the GitHub links to target files, you'll likely end up with an .html file instead of a .py or .pl file.

Step #4) Combine, remove duplicates, and re-index your .bam files

4a) Open an interactive Docker session, with access to the folder containing your .bam file (or the folder containing these scripts). For example, on Windows, your command will likely look something like:

docker run -it -v /c/Users/Charles/Documents/DNAseq_templates/My_Veritas_WGS:/mnt/wgs cwarden45/dnaseq-dependencies

4b) Move the folder containing the .bam file (run cd /mnt/wgs)

4c) If the scripts and BAM folder (named chr_bam) are in the same directory, run python combine_bams.py. Otherwise, you can specify the full file path using python combine_bams.py --bam_folder=chr_bam

The Java memory limit is set to 4GB (4g). If you have more memory (and have allocated extra memory/CPU to docker), you can change this setting via python combine_bams.py --java_mem=4g. You can also decrease the memory requirements, but the script already takes a few hours to run (with default settings) on my PC with 8 GB RAM and 4 CPU (with 5 GB RAM and 4 CPU allocated to the Docker VM). Run python combine_bams.py --help for more information.

If everything works correctly, you can delete the veritas_wgs.bam file after the script stops running.

Now, you can visualize coverage for all chromosomes simultaneously in IGV. After loading veritas_wgs.sort.filtered.bam in IGV, go to Tools --> igvtools --> count (produces veritas_wgs.sort.filter.bam.tdf file). You might want to increase the window size to produce the .tdf file more quickly (perhaps try 200+ bp), but I was able to keep the 25 bp window without any problems.

Step #5) Produce snpEff variant summary

5a) You can produce an HMTL summary of functional annotations for your variants (within the Docker container) by running python run_SnpEff.py --input=[ID].vcf. The HTML file will be called "snpEff_summary.html". It might take a few hours to produce this file.

5b) If you run the scripts in the "Advanced Users" section, you can also use snpEff to add gene annotations to your structural variant calls.

For Advanced Users

1) Use ANNOVAR for additional annotations

You can download ANNOVAR here. It is free for private use, but you will be encouraged to buy a commerical license if you provide a .com e-mail address. So, it would be best if you could provide an educational or non-profit e-mail address.

The script assumes that the ANNOVAR code as been extracted in the same directory as your .vcf file (in a folder called "annovar"). Assuming that is the case, you can use python run_ANNOVAR.py --vcf=[ID].vcf to annotate your variants.

You can see the ANNOVAR ClinVar annotations in annovar_[file name].hg19_clinvar_20160302_dropped.

2) Make additional structural variant calls using DELLY (via python run_DELLY.py), LUMPY (via python run_LUMPY.py), GASVPro (via python run_GASVPro.py), and/or Manta (via python run_Manta.py).

  • In order to run DELLY, you must first run python download_ref.py. You will be asked to enter your e-mail address as a password.
  • Delly is Dockerized, so you can download and run the Docker container (used with the script) with a command similar to this one:
docker run -it -v /c/Users/Charles/Documents/DNAseq_templates/My_Veritas_WGS:/mnt/wgs trausch/delly
  • You will probably need to allow a few hours for these programs to run.
  • You can get an idea about how the size of the deletions/duplications/insertions compare for DELLY/LUMPY versus those provided in the Vertias .vcf file by running:
python vcf_to_bed.py --vcf=[file.vcf] --caller=[Veritas/LUMPY/DELLY_DEL/DELLY_DUP/DELLY_INS]

Run python vcf_to_bed.py --help for more information.

  • You should probably also filter putative deletions based upon coverage in order to reduce false positives. I've provided a script to accomlish this:
python DEL_cov_filter.py --bed=[LUMPY_DEL.bed|DELLY_DEL.bed|GASVPro_DEL.bed|Manta_DEL.bed]

The same logic would apply to duplications, but I'm focusing on deletions on the assumption that they will be more likely to be deleterious (if they overlap a substantial portion of a genes - if you have more than a handful of filtered hits, you can use run_SnpEff.py to find exonic overlaps).

3) I've also provided scripts for analyzing 23andMe data in the 23andMe folder, which I will compare to my WGS variants. See README.md in 23andMe folder.

  • You can create .fastq files and re-align using BWA-MEM using python realign_BWA_MEM.py
  • Then, you can call additional variants using GATK (with and without an extra filter to exclude soft-clipped bases from variant calling) and/or VarScan using python run_GATK_VarScan.py.
    • VarScan will work on original alignment file, and (at least for me) original vcf was produced using freebayes. You can produce an alignment that works with all variant callers (and from which you can compare structural variant calls) using python realign_BWA_MEM.py. You can call freebayes variants on the new alignment via python run_freebayes.py, but it is not currently installed on the Docker image.
    • Depending on your computer (and what is run for variant calling), this could potentially take a few days for WGS data.

About

Scripts to Analyze DTC Sequencing and Genotyping Data (and some comparisons to Veritas WGS data)

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages