Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Computer crashed while running script_rpn_pedestrian_VGG16_caltech.m for training rpn #29

Closed
YangXiangChen opened this issue Dec 15, 2016 · 2 comments

Comments

@YangXiangChen
Copy link

Hello, I'm sorry to disturb you. I'm trying to re-implement the code and I have some questions for compiling and training the RPN.

For the compiling part, I'm not sure if I'm doing it right. First, I downloaded and compiled the Caffe, which is mentiond in the Requirements section. Second, I run the rpn_bf_build.m and startup.m and also the script_rpn_bf_pedestrian_VGG16_caltech_demo.m. The demo seems succeed, but I'm wondering is that right or not for using rpn_bf_build.m instead of faster_rcnn_build.m.

For training the RPN part, I have successfully extract images and create the annos. I also downloaded the pre-trained model from website, and unzip in the repo folder. But I have a problem on running the script_rpn_pedestrian_VGG16_caltech.m, which is that my computer will crash without warning at "preparing data" part after running for 10 or 20 minutes. My environment is ubuntu 14.04 and Nvidia Titan X, with cuda 7.5 and cudnn v3. I have noticed that even the code shows "use GPU 1", which is Titan X, the loading of my GPU memory won't increase(I use the "nvidia-smi" command in terminal to see the memory usage), but the RAM loading(system monitor) of my pc will increase and then it crashed.

Could you please give some advice? Thanks a lot!

@zhangliliang
Copy link
Owner

Hi,

The problem might be that the RPN need to pre-compute all the anchors for all the images in the dataset, and then storage them in the memory.

Actually, my computer has 64G memory, thus it's not a problem for me.

Thus, you might try to reduce the training data from Caltech 10x to Caltech 3x via modifying the parameters in the extract_img_anno.m.

Also, changeing parfor to for in the "preparing data" step could also save memory usage.

@YangXiangChen
Copy link
Author

@zhangliliang Thanks for your advice! I will give a try.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants