fMRIprep is a functional magnetic resonance image pre-processing pipeline that is designed to provide an easily accessible, state-of-the-art interface that is robust to differences in scan acquisition protocols and that requires minimal user input, while providing easily interpretable and comprehensive error and output reporting.
Python Smarty
Latest commit 5be2d4a Feb 23, 2017 @chrisfilo chrisfilo committed on GitHub Merge pull request #370 from chrisfilo/fix/hmc_interpolation
HMC refactor

README.rst

fmriprep: A Robust Preprocessing Pipeline for fMRI Data

This pipeline is developed by the Poldrack lab at Stanford University for use at the Center for Reproducible Neuroscience (CRN), as well as for open-source software distribution.

https://circleci.com/gh/poldracklab/fmriprep/tree/master.svg?style=shield Documentation Status Latest Version

About

fmriprep is a functional magnetic resonance imaging (fMRI) data preprocessing pipeline that is designed to provide an easily accessible, state-of-the-art interface that is robust to differences in scan acquisition protocols and that requires minimal user input, while providing easily interpretable and comprehensive error and output reporting. It performs basic processing steps (coregistration, normalization, unwarping, noise component extraction, segmentation, skullstripping etc.) providing outputs that make running a variety of group level analyses (task based or resting state fMRI, graph theory measures, surface or volume, etc.) easy.

Note

fmriprep performs minimal preprocessing. Here we define 'minimal preprocessing' as motion correction, field unwarping, normalization, field bias correction, and brain extraction. See the workflows_ for more details.

The fmriprep pipeline primarily utilizes FSL tools, but also utilizes ANTs tools at several stages such as skull stripping and template registration. This pipeline was designed to provide the best software implementation for each state of preprocessing, and will be updated as newer and better neuroimaging software become available.

This tool allows you to easily do the following:

  • Take fMRI data from raw to full preprocessed form.
  • Implement tools from different software packages.
  • Achieve optimal data processing quality by using the best tools available.
  • Generate preprocessing quality reports, with which the user can easily identify outliers.
  • Receive verbose output concerning the stage of preprocessing for each subject, including meaningful errors.
  • Automate and parallelize processing steps, which provides a significant speed-up from typical linear, manual processing.

More information and documentation can be found here:

https://fmriprep.readthedocs.io/

Principles

fmriprep is built around three principles:

  1. Robustness - the pipeline adapts the preprocessing steps depending on the input dataset and should provide results as good as possible independently of scanner make, scanning parameters or presence of additional correction scans (such as fieldmaps)
  2. Ease of use - thanks to dependance on the BIDS standard manual parameter input is reduced to a minimum allow the pipeline to run in an automatic fashion.
  3. "Glass box" philosophy - automation should not mean that one should not visually inspect the results or understand the methods. Thus fmriprep provides for each subject visual reports detailing the accuracy of the most importatnt processing steps. This combined with the documentation can help researchers to understand the process and decide which subjects should be kept for the group level analysis.

Acknowledgements

Please acknowledge this work mentioning explicitly the name of this software (fmriprep) and the version, along with the link to the GitHub repository (https://github.com/poldracklab/fmriprep).

License information

We use the 3-clause BSD license; the full license is in the file LICENSE in the fmriprep distribution.

All trademarks referenced herein are property of their respective holders.

Copyright (c) 2015-2017, the fmriprep developers and the CRN. All rights reserved.