Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Wikikgv2 Model Training is Hanging #11

Closed
HarryShomer opened this issue Oct 4, 2022 · 15 comments
Closed

Wikikgv2 Model Training is Hanging #11

HarryShomer opened this issue Oct 4, 2022 · 15 comments

Comments

@HarryShomer
Copy link

HarryShomer commented Oct 4, 2022

Hi,

I'm having an issue where the model gets stuck while training. It typically happens early in the first epoch. Below is an example when running the smore/training/vec_scripts/train_shallow_wikikgv2.sh (unmodified sans the GPUs) on 4 NVIDIA RTX A6000 50GB GPUs.

model_stuck

It hangs forever unless I stop it with a keyboard interrupt. Doing so yields the following traceback (I only post a portion because it's very long and repetitive).

model_traceback

It seems like something is happening in the multiprocessing as it's hanging when sharing messages between processes.

Any help would be appreciated! @hyren

Thanks,
Harry

@hyren
Copy link
Collaborator

hyren commented Oct 9, 2022

Hi, this does not seem to be related to the smore repo. Here is an instruction I found, can you please check?
https://www.rdmamojo.com/2012/05/18/libibverbs/

@HarryShomer
Copy link
Author

That doesn't seem to be the issue.

I tried seting up libibverbs, and while the warnings went away, the code is still hanging.

smore_hang

@hyren
Copy link
Collaborator

hyren commented Oct 11, 2022

Hi, can you say more about the details of the environment?

@HarryShomer
Copy link
Author

Sure. Here's some basic info. Let me know if you would like anything else.

OS: Ubuntu 20.04.5 LTS
CUDA: 11.6.124
GPU(s): NVIDIA RTX A6000
Python: 3.9.12
PyTorch: 1.12.1

@Hanjun-Dai
Copy link
Collaborator

Hi there, sorry for the inconvenience, could you please quickly try with only 1 gpu, and see if it still hangs? We are trying to figure out whether it is due to gpu-gpu communication, or the c++ sampler we used.

@HarryShomer
Copy link
Author

Sorry, I should have been clearer in my initial comment. The code doesn't hang when running with just one GPU. This only occurs when using multiple GPUs.

@Juanhui28
Copy link

We tried one gpu again, actually it still hangs. Sorry for the inconvenience.

@Hanjun-Dai
Copy link
Collaborator

Hi there,

Could you please kindly pull the latest code in the wikikgv2 branch, and add --train_async_rw=False to your script and try again? Since I'm unable to reproduce your issue, I'd like to see whether this would temporarily resolve the issue.

@Juanhui28
Copy link

Hi,

We tried your suggestion and now the training doesn't hang! But we got the similar issue in the evaluation. And we noticed your suggestion in another issue and follow the suggestion, but it still hangs.

I really appreciate your help!

@hyren
Copy link
Collaborator

hyren commented Oct 12, 2022

hi, what's the script you used?

@Juanhui28
Copy link

Hi, train_shallow_wikikgv2.sh in the training/vec_scripts folder.

@hyren
Copy link
Collaborator

hyren commented Oct 12, 2022

Just to make sure, you used the scripts in train_shallow_wikikgv2.sh and add the --train_async_rw=False flag?

@Juanhui28
Copy link

Juanhui28 commented Oct 12, 2022

Actually no, since if we add it we got the unrecognized arguments error.

@hyren
Copy link
Collaborator

hyren commented Oct 12, 2022

have you pulled the most recent commits?

@Juanhui28
Copy link

Sorry we made some mistakes when we merged the code. Now we pulled the the latest code and it works for both training and evaluation!

Thank you so much!

@hyren hyren closed this as completed Oct 13, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants