Benchmark Models for WebVision 2.0 Dataset
Please check the 2020 Webvision dev kit here
Step-1: Prepare data
The data directory should be prepared following the structure stated in this link. We have provided a download + preprocess script for this.
cd utils bash download.sh
Note that the server hosting Webvision Data reboots every day at midnight (Zurich time). You might want to change wget to something else.
Step-2: Install tensorpack
Following the instructions here to install tensorpack.
pip3 install tensorpack==0.9.1
Step-3: Train the model (ResNet-50)
Following the setting for ImageNet in tensorpack, we use 4 GPUS with the batch size being set to 64x4=256. Run the following script to train the model,
python3 ./imagenet-resnet.py --data /raid/webvision2018/ --gpu 0,1,2,3 -d 50 --mode resnet
We offer several pretrained models. Due to the class imbalance in WebVision, we duplicated the file items in train.txt such that different classes have equal number of training samples. You might want to add similar strategies in imagenet5k.py or modify your own train.txt. Check utils/upsample.py for an example.
|ResNet-50 (101 Epoch)||54.28%||30.69%||link|
|ResNet-50 (205 Epoch)||52.10%||28.51%||link|
|ResNet-101 (100 Epoch)||52.21%||28.62%||link|
|ResNet-101 (200 Epoch)||50.12%||26.78%||link|
|ResNet-101 (300 Epoch)||48.97%||25.74%||link|
|ResNet-101 (500 Epoch)||48.38%||25.21%||link|
|ResNeXt-101 (100 Epoch)||50.62%||27.11%||link|
|ResNet-152 (100 Epoch)||51.23%||27.80%||link|
|ResNet-152 (200 Epoch)||48.98%||25.75%||link|
|ResNet-152 (300 Epoch)||48.05%||24.88%||link|
|ResNet-152 (500 Epoch)||47.31%||24.31%||link|
|ResNet-152-SE (100 Epoch)||51.61%||28.02%||link|
Note that here one epoch is defined as an iteration over 1280000 images (ImageNet Epoch), you could divide it by 16M/1.28M to get the Webvision epoch count. Scores are based on validation set, and calculated by utils/eval_score.py using the same metric (Equal weight for each class) as the competition evaluation.
All models are trained from scratch using solely Webvision data.
Evaluation and Submission
To generate the prediction files for CodaLab submissions, assume testimages are stored in the above format in /raid/webvision2018_test/val/:
python3 ./imagenet-resnet.py --data /raid/webvision2018_test/ -d 50 --mode resnet --eval --load train_log/imagenet-resnet-d50-webvision2018-200epochs/model-1055000 # Preparing the submission file python3 utils/json2sub.py
Note that the error rate in the above figure is calculated directly over validation set (without averaging over class), and thus not the same metric used by the evaluation system.
The code is adapted from Tensorpack, where we modified the data pipeline for Webvision.
The webvision dataset is collected by the learning from web data team at Computer Vision Laboratory, ETH Zurich. The dataset development was supported by Google Research Zuirch. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the funding agencies.
This repository is developed by the learning from web data team at Computer Vision Laboratory, ETH Zurich. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the funding agencies.