New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
System memory usage increase in training #12
Comments
Hi, I have never observed this issue, but it is also true that I always had more than 16gb memory. Can you try to somehow find where the memory leak is or try to train it on a machine with 32gb ram? For the other issue that you posted: I have uploaded the preprocesing scripts that we have used for semantic_kitti and other datasets |
Thank you very much @zgojcic ! I will look into the possible memory leaks, and thanks for the preprocesing scripts. For the memory issue, |
After some more time searching on the web, I found this that could be a possible explanation. Maybe it has to do with the dataloader iterating across lists and dicts which adds up over time? The suggested solution is to replace them with numpy arrays. |
Hi @Alt216 this could indeed be the case, at the moment I do not have time to investigate this (especially as it work ok on machines with more RAM), but if you can find the solution it would be great if you can make a PR. Best |
Closing due to inactivity. |
Hi, when I run
python train.py ./configs/train/train_weakly_supervised.yaml
to train the network from scratch using our dataset, my system memory usage will slowly increase until it max out the system memory and then the traning will crash. I have 16gb of system memory and the training can only go on for a little more than one epoch with ~16000 training samples. I tried to lower thenum_workers
to 4 and lower the batch size to 2 but they didn't seem to resolve the issue.The text was updated successfully, but these errors were encountered: