[D] Neat project that would "fit" onto a 4090? Submitted by lifesthateasy t3_11ktxjl on March 7, 2023 at 8:49 AM in MachineLearning 8 comments 10
gmork_13 t1_jbbj49n wrote on March 7, 2023 at 9:19 PM With fp16/int8 you can probably stick a couple of LLMs of smaller size onto that card. Have a look around, with fp32 it's about 1B params per 4GB of VRAM. Halve it for fp16 and again for int8 (very roughly). Permalink 2
Viewing a single comment thread. View all comments