AiChip t1_j74ku5a wrote
Wow! This is huge! 1B parameters model beating 175 B parameters model…
Lengador t1_j74ro7q wrote
That's the number in the headline, but if you look at the tables you can see their 223M parameter model beats the 175B parameter model significantly as well. That's 0.1% the size! Absolutely insane.
HeyLittleTrain t1_j77w36w wrote
At what size could I run a model on a decent gaming PC?
emotionalfool123 t1_j78bjj2 wrote
Stable diffusion is around 866M params which can be run on 12gb 3080
7734128 t1_j9j9r06 wrote
And on my 8 GB GTX 1080.
Lengador t1_j78ovy2 wrote
You can (just) run a 1B parameter model on a good gaming rig.
i2mi t1_j786bu0 wrote
Around 2M Edit: the number I gave is completely delusional. Sorry
HeyLittleTrain t1_j7avkil wrote
Your answer seems substantially different than the others.
NapkinsOnMyAnkle t1_j9jtolb wrote
I've trained 100m CNNs on my laptop 3070 6gb. So...
[deleted] t1_j77e3ku wrote
[deleted]
JClub t1_jabyi76 wrote
GPT was never trained with image data, why is this a fair comparison? The UnifiedQA model is from 2022, so it doesn't seem fair either. Why don't we have some comparisons with other SOTA multimodal models? Such as OFA or UniT
Viewing a single comment thread. View all comments