You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardexpand all lines: README.md
+22-49
Original file line number
Diff line number
Diff line change
@@ -5,6 +5,8 @@ This is a PyTorch implementation of semantic segmentation models on MIT ADE20K s
5
5
ADE20K is the largest open source dataset for semantic segmentation and scene parsing, released by MIT Computer Vision team. Follow the link below to find the repository for our dataset and implementations on Caffe and Torch7:
6
6
https://github.com/CSAILVision/sceneparsing
7
7
8
+
If you simply want to play with our demo, please try this link: http://scenesegmentation.csail.mit.edu You can upload your own photo and segment it!
The training is benchmarked on a server with 8 NVIDIA Pascal Titan Xp GPUs (12GB GPU memory), ***except for*** ResNet101dilated, which is benchmarked on a server with 8 NVIDIA Tesla P40 GPUS (22GB GPU memory), because of the insufficient memory issue when using dilated conv on a very deep network. The inference speed is benchmarked a single NVIDIA Pascal Titan Xp GPU, without visualization.
@@ -158,6 +155,7 @@ The training is benchmarked on a server with 8 NVIDIA Pascal Titan Xp GPUs (12GB
158
155
The code is developed under the following configurations.
159
156
- Hardware: 1-8 GPUs (with at least 12G GPU memories) (change ```[--gpus GPUS]``` accordingly)
## Quick start: Test on an image using our trained model
163
161
1. Here is a simple demo to do inference on a single image:
@@ -167,86 +165,61 @@ chmod +x demo_test.sh
167
165
```
168
166
This script downloads a trained model (ResNet50dilated + PPM_deepsup) and a test image, runs the test script, and saves predicted segmentation (.png) to the working directory.
169
167
170
-
2. To test on multiple images or a folder of images, you can simply do something as the following (```$PATH_IMG1, $PATH_IMG2, $PATH_IMG3```are your image paths):
168
+
2. To test on an image or a folder of images (```$PATH_IMG```), you can simply do the following:
3. See full input arguments via ```python3 test.py -h```.
180
-
181
173
## Training
182
174
1. Download the ADE20K scene parsing dataset:
183
175
```bash
184
176
chmod +x download_ADE20K.sh
185
177
./download_ADE20K.sh
186
178
```
187
-
2. Train a model (default: ResNet50dilated + PPM_deepsup). During training, checkpoints will be saved in folder ```ckpt```.
179
+
2. Train a model by selecting the GPUs (```$GPUS```) and configuration file (```$CFG```) to use. During training, checkpoints by default are saved in folder ```ckpt```.
188
180
```bash
189
-
python3 train.py --gpus GPUS
181
+
python3 train.py --gpus $GPUS --cfg $CFG
190
182
```
191
-
192
183
- To choose which gpus to use, you can either do ```--gpus 0-7```, or ```--gpus 0,2,4,6```.
193
184
194
-
For example:
185
+
For example, you can start with our provided configurations:
3.See full input arguments via ```python3 train.py -h```.
202
+
3.You can also override options in commandline, for example ```python3 train.py TRAIN.num_epoch 10```.
218
203
219
204
220
205
## Evaluation
221
-
1. Evaluate a trained model on the validation set. ```--id``` is the folder name under ```ckpt``` directory. ```--suffix``` defines which checkpoint to use, for example ```_epoch_20.pth```. Add ```--visualize``` option to output visualizations as shown in teaser.
0 commit comments