Course project for IE529: Stats of Big data & Clustering, 2017 Fall, UIUC
Name | Guthub Homepage |
---|---|
Jvn Karthik | N/A |
Naman Shukla | namanUIUC |
Shubham Bansal | bansalshubh91 |
Zhenye Na | Zhenye-Na |
Ziyu Zhou | Ziyu0 |
We implemnt the experiments presented in the paper Nonlinear Component Analysis as a Kernel Eigenvalue Problem by Bernhard Schölkopf, Alexander Smola, Klaus-Robert Müller. Also, we write our own example on Kernel PCA. In this regard, you can read our report and our presentation slides.
In order to run the experiments, make sure you have all dependencies installed
- matplotlib (>= 2.0.0)
- scipy (>=0.19.0)
- numpy (>=1.12.1)
- sklearn (>=0.0)
You can install them by typing
pip3 install 'whatever you need'
The prgramming languages we use are Python and MATLAB. If you do not have access to MATLAB on your laptop. We advise you to install Octave
instead. You can refer to this webpage for installing.
In the paper, there are two major experiments:
- Toy example: 4-degree Polynomial Kernel PCA
- Character Recognition (USPS Dataset)
We implemented this part with MATLAB. The code can be found here.
Principal Component Analysis (PCA) is a dimensionality reduction technique that is used to transform and a high-dimensional dataset into a smaller dimen- sional subspace to give a directed impression of the dataset prior to running a machine learning algorithm on the data. The Iris dataset is in a 4 th dimensions (features) of three different iris ower species.
Related codes can be found here.
The dataset contains numeric data obtained from the scanning of handwritten digits from envelopes by the U.S. Postal Service. The original scanned digits are binary and of different sizes and orientations; the images here have been deslanted and size normalized, resulting in 16 × 16 grayscale. We will first extract features via Kernel PCA and apply that to a SVM classifier to train and test on the splitted USPS dataset.
Related codes can be found here.