AttributeError: module 'torch.nn.functional' has no attribute 'scaled_dot_product_attention' #2260
Closed
1 task done
Labels
bug
Something isn't working
Describe the bug
I would like to use the TheBloke/Wizard-Vicuna-13B-Uncensored-GPTQ model on my RTX3080 with 10GB of VRAM.
It works great at first but once the input gets too big, it does not work anymore because of not enough VRAM.
So I'm trying to use the model with
pre_layer
set to 20 to move a part of the model to the CPU, but when I do that I receive the following error:My version of pytorch is 1.13.1 in that Python environment, and I believe that the
scaled_dot_product_attention
attribute is only available with a newer version (2.0) of pytorch.Note that I had to change the GPTQ-for-LLaMa repository in repositories/GPTQ-for-LLaMa to the a recent commit of the qwopqwop200 repo to make this model work.
What should I do? Is it safe to try to upgrade to pytorch 2.0 or will it causes other issues.
Is there an existing issue for this?
Reproduction
Try to use the above model with pre_layer set to 20 and the GPTQ-for-LLaMa repo set the the commit above.
Screenshot
No response
Logs
System Info
The text was updated successfully, but these errors were encountered: