Submitted by Nerveregenerator t3_z0msvy in deeplearning
Nerveregenerator OP t1_ix75olf wrote
Reply to comment by scraper01 in GPU QUESTION by Nerveregenerator
So I did some research. According to the lambda labs website, 4 1080s combined will get me 1.5x throughout as a 3090 with FP32 training. FP16 seems to yield a 1.5x speed up for the 3090 for training. So even with mixed precision, it comes out to be the same. The actual configuration of 4 cards is not something I’m very familiar with, but I wanted to point this out as it seems like NVIDIA has really bullshitted a lot with their marketing. A lot of the numbers they throw around just don’t translate to ML.
incrediblediy t1_ix7czdr wrote
> 4 1080s combined will get me 1.5x throughout as a 3090 with FP32 training. FP16 seems to yield a 1.5x speed up for the 3090 for training.
I think that's when only comparing CUDA cores without Tensor cores, anyway you can't merge VRAM together for large models
Viewing a single comment thread. View all comments