open-source feature selection repository in python (DMML Lab@ASU)
Switch branches/tags
Nothing to show
Clone or download
Latest commit 2278f3f Oct 9, 2017
Failed to load latest commit information.
skfeature update Oct 8, 2017
.gitignore Initial commit Nov 6, 2015
LICENSE Initial commit Nov 6, 2015
PKG-INFO update Mar 9, 2016 Revert "Python 2 OR Python 3 can be used" Oct 8, 2017 update Mar 9, 2016


Feature selection repository scikit-feature in Python (DMML Lab@ASU).

scikit-feature is an open-source feature selection repository in Python developed by Data Mining and Machine Learning Lab at Arizona State University. It is built upon one widely used machine learning package scikit-learn and two scientific computing packages Numpy and Scipy. scikit-feature contains around 40 popular feature selection algorithms, including traditional feature selection algorithms and some structural and streaming feature selection algorithms.

It serves as a platform for facilitating feature selection application, research and comparative study. It is designed to share widely used feature selection algorithms developed in the feature selection research, and offer convenience for researchers and practitioners to perform empirical evaluation in developing new feature selection algorithms.

Installing scikit-feature


Python 2.7 and Python 3





After you download from the project website (, unzip the file.

For Linux users, you can install the repository by the following command:

python install

For Windows users, you can also install the repository by the following command: install

Project website

Instructions of using this repository can be found in our project webpage at


If you find scikit-feature feature selection reposoitory useful in your research, please consider citing the following paper::

   title={Feature Selection: A Data Perspective},
   author={Li, Jundong and Cheng, Kewei and Wang, Suhang and Morstatter, Fred and Trevino, Robert P and Tang, Jiliang and Liu, Huan},
   journal={arXiv preprint arXiv:1601.07996},


Jundong Li E-mail:

Kewei Cheng E-mail: