-
Notifications
You must be signed in to change notification settings - Fork 57
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Error in Self-Supervised Instruction Tuning #17
Comments
excuse me, have u solved this problem? |
Thank you for your interest in our GraphGPT. I apologize for the delayed response due to the academic workload at the end of the semester.
If that doesn't work, feel free to ask me further. |
Thank you for your attention. Please refer to my reply above. |
We ran into the same problem, although we commented out "replace_llama_attn_with_flash_attn()", but still got an error: `['model.embed_tokens.weight', 'model.graph_projector.weight', 'model.graph_projector.bias'] 0%| | 0/137400 [02:11<?, ?it/s] ` |
Hi there, thanks for offering this interesting project! I have trouble when conducting the Self-Supervised Instruction Tuning. Specifically, the error goes as follows:
I use the suggested configurations (environments, scripts) and conduct the tuning on a Linux server equipped with 4 A100 in a distributed manner. Still, I have also tried to conduct the tuning on one GPU merely. To avoid CUDA OOM error, I have modified the train/eval batch size to 1. However, I have encountered another error as follows:
Therefore, the tuning process can not be reproduced on either single or multiple GPUs. Any suggestions for troubleshooting would be appreciated. Looking forward to your kind reply!
The text was updated successfully, but these errors were encountered: