Which affordable GPU would you recommend to train a language model?
It is diffucult to say which GPU fits best for your usecase since larger language models typically require more computational power and VRAM than smaller ones. A recommendation would be the 24GB version of the NVIDIA Tesla M40 since it has a lot of VRAM and can be bought used on sites like https://ebay.com/ for below 150$, however it is worth mentioning that this GPU was launched all the way back in November 10th, 2015 so you shouldnt expect groundbreaking performance. If your budged is a little higher, I could also recommend the NVIDIA Tesla K80 24GB.
I want to finetune GPT-J-6B, would those be good enough for that?