Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Inference time question #141

Closed
huang45 opened this issue Sep 22, 2020 · 5 comments
Closed

Inference time question #141

huang45 opened this issue Sep 22, 2020 · 5 comments

Comments

@huang45
Copy link

huang45 commented Sep 22, 2020

Hi, what is the GPU used to test the inference time in the model zoo? I tested HRNetV2p-W18-Small model by 2080ti and only get ~0.3 seconds. Is that normal?

@York1996OutLook
Copy link

i got the same result:
out:
(open-mmlab) yaochunchun@dl-7047:~/mmsegmentation$ python tools/benchmark.py configs/fastscnn/fast_scnn_4x8_80k_lr0.12_cityscapes.py \

pretrained_models/fast_scnn_4x8_80k_lr0.12_cityscapes-cae6c46a.pth

2020-09-22 11:01:57,902 - mmseg - INFO - Loaded 500 images
Done image [50 / 200], fps: 15.14 img / s
Done image [100/ 200], fps: 15.09 img / s
Done image [150/ 200], fps: 15.08 img / s
Done image [200/ 200], fps: 15.02 img / s

and from the paper:
Input Size | Class | FPS
-- | -- |
1024 × 2048 | 68.0 | 123.5
512 × 1024 | 62.8 | 285.8
256 × 512 | 51.9 | 485.4

@huang45
Copy link
Author

huang45 commented Sep 22, 2020

tested for hrnet w18s (512*512) by benchmark.py and got following result (it looks better):

python tools/benchmark.py configs/hrnet/fcn_hr18s_512x512_160k_my.py checkpoints/fcn_hr18s_512x512_160k_my_save/iter_32000.pth
2020-09-22 14:12:18,743 - mmseg - INFO - Loaded 601 images
Done image [50 / 200], fps: 48.45 img / s
Done image [100/ 200], fps: 48.42 img / s
Done image [150/ 200], fps: 48.39 img / s
Done image [200/ 200], fps: 48.34 img / s
Overall fps: 48.34 img / s

but ~0.16 seconds was showed during the training:

2020-09-22 11:28:45,243 - mmseg - INFO - Iter [50/80000] lr: 1.199e-01, eta: 3:41:06, time: 0.166, data_time: 0.004, memory: 2135, decode.loss_seg: 0.5788, decode.acc_seg: 64.0334, aux_0.loss_seg: 0.2789, aux_0.acc_seg: 62.9780, aux_1.loss_seg: 0.3045, aux_1.acc_seg: 59.7900, loss: 1.1622

@xvjiarui
Copy link
Collaborator

Hi @huang45
We benchmarked with V100 machines. Please refer to here for hardware reference.
As for the training, the time is not accurate at the beginning of training stage. There is a lot going on, e.g. creating worker process, finding the optimial cudnn algorithm for current hardware and input, etc.

@xvjiarui
Copy link
Collaborator

Hi @York1996OutLook
Our implementation framework is not the same as the original paper. You are very welcomed to comment here if there is some better(faster) PyTorch implementation.

@huang45
Copy link
Author

huang45 commented Sep 24, 2020

@xvjiarui Got it, thank you!

Leeinsn added a commit to Leeinsn/mmsegmentation that referenced this issue Dec 12, 2022
Leeinsn added a commit to Leeinsn/mmsegmentation that referenced this issue Jan 12, 2023
Leeinsn added a commit to Leeinsn/mmsegmentation that referenced this issue Jan 12, 2023
Leeinsn added a commit to Leeinsn/mmsegmentation that referenced this issue Jan 12, 2023
MeowZheng added a commit that referenced this issue Jan 20, 2023
## Motivation

Support for biomedical 3d images augmentation.

## Modification

Add BioMedical3DRandomFlip in mmseg/datasets/transforms/transforms.py.

Co-authored-by: MeowZheng <meowzheng@outlook.com>
sibozhang pushed a commit to sibozhang/mmsegmentation that referenced this issue Mar 22, 2024
nahidnazifi87 pushed a commit to nahidnazifi87/mmsegmentation_playground that referenced this issue Apr 5, 2024
…mlab#2404)

## Motivation

Support for biomedical 3d images augmentation.

## Modification

Add BioMedical3DRandomFlip in mmseg/datasets/transforms/transforms.py.

Co-authored-by: MeowZheng <meowzheng@outlook.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants