What can a 40gb+ vram gpu train that a 24gb vram gpu can't? #912
Replies: 1 comment 8 replies
-
I think you are unlikely to get a detailed answer in the form of real experience with video cards with 48 or 80 GB of memory. People with 3090 and 3090TI are quite rare, and those who understand their capabilities in terms of training are probably not even dozens, but only a few. I can give you an example of my transition from 3060 to 3090. On 3060 I trained with fp16 and 8bit adam and without gradient_checkpointing I was able to set max batch size to 3. With gradient_checkpointing it was 10. On 3090 without gradient_checkpointing batch size I can set 20. With gradient_checkpointing set to 70. Also, if we consider the batch size, then it has a non-uniform rectilinear distribution of efficiency. The most efficient value is of course when the batch size is equal to the number of images. But the more images, the bigger the batch size. And I roughly measured that increasing the batch size value by 20 at fp16, 8 bits adam and gradient_checkpointing increases memory consumption by about 4GB. At the same time, an increase in batch size from 1, the efficiency from its value has the fastest growth rate up to 10, then the efficiency growth rate begins to fall more and more. In theory, with this amount of memory, you could run training in fp64 without any problems, only technically this is not yet implemented, since Adams do not yet support this format. Perhaps, in theory, this amount of memory will give you the opportunity to work with neural networks in terms of already vid2vid, for example, with the recently shown Gen-1 from Runway. But here again there are many questions: when will it be released? will it be publicly available? what are the requirements? etc. |
Beta Was this translation helpful? Give feedback.
-
Considering buying another gpu in the future. Either another RTX 3090 (24gb) or an RTX A6000 (48gb). But the A6000 is 5x expensive, and basically the same speed. An 80gb gpu is 15x as expensive.
Since it's for my company, I'm willing to invest if a 40+gb vram gpu can deliver a higher quality model. Higher quality output meaning: higher detailed images, higher resolution, and pictures that more accurately represent what I'm trying to create. Model training speed is not a concern. I'm only concerned with quality.
Anybody got any experience to tell me what I'll be able to do with a 40+gb vram gpu and that I can't already do with my 24gb vram gpu?
Beta Was this translation helpful? Give feedback.
All reactions