Skip to content
Repo for Mincall - MinION basecaller we're working on during academic year 2016/2017.
Jupyter Notebook Python Other
Branch: master
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Permalink
Type Name Latest commit message Commit time
Failed to load latest commit information.
mincall
notebook
.dockerignore
.gitignore
Dockerfile
Dockerfile.gpu
Dockerfile.inference.gpu
LICENSE
Makefile
README.md
nmiculinic_ma.pdf
requirements.txt
setup.cfg
setup.py

README.md

MinCall

This is MinION DNA basecaller developed from 2016-2018, and served as mine master thesis. The whole master thesis is attached to this repo with the detailed introduction, problem description, related work, methodology, results and final commentary.

For a quick summary of master thesis & this repo:

  • It's a deep learning approach using a sequence to sequence models with CTC loss with CNN's.
  • No LSTM were used which leads to high speeds.
  • During performance tuning Beam-search proved the bottleneck.
  • If the Greedy search is used per-read results are slightly worse, but the read consensus is overall better.
  • Whole code uses modern tensorflow and keras, with a custom pre-processing pipeline.
  • It uses simple YAML config file, validated using the voluptuous library
  • Hyperparameter search is implemented using pluggable architecture. Currently, SigOpt and RandomSearch strategies are supported.
  • This is primarily scientific code, thus I've aimed to the best quality and software engineering practices possible within given timeline, though priority were the results. This code has been through one rewrite from scratch which improved its quality and ease of use.
  • Many progress bars are implemented using tqdm library and it's possible to add graylog logging backend. I've used graylog for log management during training, much better than searching through log files.
  • Design goal was easy reproducibility. Thus config files were preferred over command line arguments. Hyperparameter search creates valid training config files and starts the training code. That code is no smarter whether it's part of hyperparam search or single run experiment.
  • For further questions feel free to contact me on my email or open the issue. University one is disabled (( I've graduated )) but a private one is available on my GitHub profile.

Setup

  • Install via python setup.py install or use docker images (( via makefile ))
  • python3 -m mincall basecall --in <fast5 folder> --out out.fasta --model model.save
  • Or via docker it's similar:
    • docker run --rm -it nmiculinic/mincall:latest-py3 basecall ...
    • docker run --runtime=nvidia --rm -it nmiculinic/mincall:latest-py3-gpu basecall ... GPU version required nvidia-docker installed

Training

python -m mincall train --config mincall/example/config.yml

If you're doing training I recommend checking the source code. mincall/train/_train.py is the entrypoint and should be started through there. You can also see how the config file is parsed and used. Data is formated via gzipped protobuf format defined in minion-data repository.

For all other information look at the attached master thesis (nmiculinic_ma.pdf), and the source code.

Further work

  • Perform model training on bigger dataset than E. Coli R9.4 . My estimate is the results shall be similar to other basecallers.
You can’t perform that action at this time.