Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

CUDA out of memory #28

Closed
ruodingt opened this issue Apr 16, 2020 · 3 comments
Closed

CUDA out of memory #28

ruodingt opened this issue Apr 16, 2020 · 3 comments

Comments

@ruodingt
Copy link

Hi @youngwanLEE
I was trying centermask2 on a different dataset other than COCO.
I use a single V100 GPU.

I put Batch size to 8 and remain MIN_SIZE_TRAIN unchanged.
The config file I used is centermask_V_39_eSE_FPN_ms_3x.yaml

Yet I still got CUDA OOM error.

I couldn't see other factors that could leading to this OOM error.

Could you please give me some tips?

@youngwanLEE
Copy link
Owner

@ruodingt

why don't you lower the batch size to 4?

@ruodingt
Copy link
Author

You are totally right.

In your experiment script and paper, I saw you are using 16 as batch size.
May I ask whether 16 is for per-GPU or not?

Thank you.

@jgsch
Copy link

jgsch commented Apr 17, 2020

@ruodingt > from the detectron2 documentation:

# Number of images per batch across all machines.
# If we have 16 GPUs and IMS_PER_BATCH = 32,
# each GPU will see 2 images per batch.
_C.SOLVER.IMS_PER_BATCH = 16

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants