-
Notifications
You must be signed in to change notification settings - Fork 99
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Question of training utilizing A6000 #71
Comments
When i tried to train this model, i couldn't train with A6000. |
Same issue here. |
I met same issue, if anyone has found the solution of it plz share :) |
I removed a 'replace_llama_attn_with_flash_attn()' statement from the 'video_chatgpt/train/train_mem.py' path and then the training proceeded. Could removing this code cause any issues with performance? |
I used A40 GPUs and got same issue here. How should I solve this problem? |
Hi @everyone, Flash Attention only works on A100 or H100. In case if you want to train on any other GPU, commenting out the line at
Please let me know if you will have any questions. |
Hello, first of all, I would like to express my deep gratitude for your excellent research.
I'm currently conducting training with A6000 x 8 GPUs.
But, I got below errors.
Is there a way to resolve this issue by not using flash-attention or by modifying another part of the code??
I did below train code.
The text was updated successfully, but these errors were encountered: