Demo code for Horikawa and Kamitani, "Generic decoding of seen and imagined objects using hierarchical visual features", https://www.nature.com/articles/ncomms15037.
Matlab C Python
Switch branches/tags
Nothing to show
Clone or download
Permalink
Failed to load latest commit information.
code [docs] add comments Jun 5, 2018
data Add prototxt files for CNN nets Jul 13, 2018
docs Set theme jekyll-theme-cayman Apr 26, 2018
.gitignore Add .gitignore Feb 21, 2018
README.md Update README.md Apr 2, 2018

README.md

Generic Object Decoding

This repository contains the data and demo codes for replicating results in our paper: Horikawa, Kamitani "Generic decoding of seen and imagined objects using hierarchical visual features". The generic object decoding approach enabled decoding of arbitrary object categories including those not used in model training. For more technical details, please refer to the paper: https://www.nature.com/articles/ncomms15037.

Data (fMRI data and visual features)

The preprocessed fMRI data for five subjects (training, test_perception, and test_imagery) and visual features (CNN1-8, HMAX1-3, GIST, and SIFT) are available at http://brainliner.jp/data/brainliner/Generic_Object_Decoding. The fMRI data were saved as the BrainDecoderToolbox2 format (https://github.com/KamitaniLab/BrainDecoderToolbox2). The unpreprocessed fMRI data is available from the OpenNeuro (https://openneuro.org/datasets/ds001246).

Visual images

For copyright reasons, we do not make the visual images used in our experiments publicly available on our server. You can directly download the images from URLs listed in the imageURL_training.csv (1-1200 training images) and imageURL_test.csv (1-50 test images) files.

While most of the images can be downloaded from the URLs, some of the images have been deleted from the original URLs. In that case, you need to create your account for the ImageNet (http://image-net.org/index) to download the image set for each synset (2011, fall release; http://image-net.org/download-images), and search corresponding images using the synset ID and image ID listed in imageID_training.csv and imageID_test.csv files. The image filename is formatted as xxxx_yyyy, where xxxx represents the WordNet synset ID and yyyy represents the image ID.

For easier access to the stimulus images used in our study, we provide a matlab code (get_stimulusimages.m) for downloading all the training and test images from the ImageNet (while you still need to register to the ImageNet). The code requires your username and accesskey, which are provided from the ImageNet, as its arguments. Using this code, you can automatically collect the images cropped and resized as in our study. Note that it will take about one day for collecting all the experimental stimuli with a single computer. But you can run the code with multiple computers in parallell to accelarate the processings.

Please contact us (kamitanilab.contact@gmail.com) if you have trouble with the above method, so that we will help you obtaining the images.

Demo program

Demo programs for Matlab and Python are available in code/matlab and code/python, respectively. See README.md in each directory for the details.