-
-
Notifications
You must be signed in to change notification settings - Fork 15
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Is an 8x NVIDIA RTX 3090 GPU Setup Sufficient for Training Models in the Predictive World Model 2024 Competition? #17
Comments
Hi, You can try using smaller image scale, set supervise_all_future to False, larger voxel size for reducing number of points for supervision, other backbones, or with_cp to True in backbone to reduce memory cost. Hope these work for you. ViDAR indeed requires a large GPU memory cost and OpenScene dataset has 8 input cameras. Thanks, |
In the past few days, I have been trying to use a single 3090 running vidar to reproduce on nuscenes. How can I reduce the memory of the nuscenes data set? |
In the past few days, I have been trying to use a single 3090 running vidar to reproduce on nuscenes. How can I reduce the memory of the nuscenes data set? |
|
Great to hear. I think reducing the computational & GPU memory cost of ViDAR is of great importance :_). |
I am excited about participating in the Predictive World Model 2024 competition and have been preparing my environment accordingly. My current setup includes a system with 8 NVIDIA RTX 3090 GPUs, which I believed would be more than capable of handling the training demands of the competition's models.
However, even after adjusting the configuration settings to the minimum requirements as per the competition guidelines, I'm encountering a persistent issue where I run out of memory. The error I receive is as follows:
RuntimeError: CUDA out of memory. Tried to allocate 60.00 MiB (GPU 7; 23.70 GiB total capacity; 21.79 GiB already allocated; 18.81 MiB free; 21.97 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
Is an 8x RTX 3090 GPU setup insufficient for training the competition models, or might there be an issue with my configuration or approach?
The text was updated successfully, but these errors were encountered: