OPT Example #1356
-
Hi all, What single GPU was used for the fine-tuning example of OPT? I am running the bash script provided in the repository with the default, 6.7b, configuration on an RTX 3090 locally, before moving to a cloud instance, but I am receiving a CUDA Memory error. The 2.7b OPT model throws a CUDA Memory error as well. The 1.3b model seems to work with a batch size of 8. Also, is there a specific reason for using I appreciate the help. Thank you, Enrico |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 4 replies
-
Hi, Enrico. First, could you offer a detailed log or more information about your CUDA Memory error. Second, we are going to update the OPT example by using our new ColoTensor API. For better perfermance and robustness, I suggest to try the example script which will be released soon. |
Beta Was this translation helpful? Give feedback.
Hi, Enrico.
First, could you offer a detailed log or more information about your CUDA Memory error.
Second, we are going to update the OPT example by using our new ColoTensor API. For better perfermance and robustness, I suggest to try the example script which will be released soon.