New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
CPU memory usage is too high and other queries #37
Comments
Thanks. It is because the data loader would load image features to memory 4 times (i.e., 4 copies of the image features are in memory). The |
Thanks for replying.
|
|
Could you suggest a better way of loading features ? I'm not able to fit them even with |
May I ask how large your main memory is?
If it still exceeds the memory limitation, the code might need to load features from disk. |
I have 8 cores in my GCP instance (around 48 GB). It just tried with with |
Thanks. WIth 48 GB main memory, it should work fine with Loading features from disk is definitely possible; Multiple workers should be involved to balance the loading. The current choice loads from memory thus ultimately remove the cost of memory loading but would be inefficient when the memory is not big enough. |
Thanks for your reply. In your experiment, did you try with other configurations (9,6,6 for L,X,R ) ? |
I did not try them given the computational resources. |
Thanks.
|
|
Thanks for replying. Sorry, I just wanted to learn more about the pretraining and I don't consider it problematic. The reason why I asked is that pretraining isn't feasible for me right now.
|
|
Hi, |
Thanks, the results of test-dev and test-std require using the test servers. The detailed processes for each dataset are provided at the end of each section. E.g., https://github.com/airsplay/lxmert#submitted-to-vqa-test-server, |
Hi, |
Currently, I did not find a clean way to fetch the attention graphs thus the code is badly organized. I just gathered all the output, saved them in tsv files, and visualize the output by ipynb. So for now, I do not have a plan to release them. |
How did you gathered the output, if you could please point the line in your current codebase, where you got the output from, that'd be really helpful. |
My way is simple but not elegant. I create a global list and append the output here to the list. The list is cleared before each forward operation and logged after the forward. |
Thanks for sharing this code. When I'm performing finetuning with VQA, my RAM usage blows up. With
num_workers
set to 4, it requires 207 GB. I've tried with different batch sizes also. The script with--tiny
flag runs successfully. But when I'm loading bothtrain and nominival
, the memory usage blows up. I get memory can't be allocated. Do you know a workaround for this ? I think this is because we are storing all the features from faster_rcnn in RAM ?The text was updated successfully, but these errors were encountered: