Lengador t1_j74ro7q wrote
Reply to comment by AiChip in [R] Multimodal Chain-of-Thought Reasoning in Language Models - Amazon Web Services Zhuosheng Zhang et al - Outperforms GPT-3.5 by 16% (75%->91%) and surpasses human performance on ScienceQA while having less than 1B params! by Singularian2501
That's the number in the headline, but if you look at the tables you can see their 223M parameter model beats the 175B parameter model significantly as well. That's 0.1% the size! Absolutely insane.
HeyLittleTrain t1_j77w36w wrote
At what size could I run a model on a decent gaming PC?
emotionalfool123 t1_j78bjj2 wrote
Stable diffusion is around 866M params which can be run on 12gb 3080
7734128 t1_j9j9r06 wrote
And on my 8 GB GTX 1080.
Lengador t1_j78ovy2 wrote
You can (just) run a 1B parameter model on a good gaming rig.
i2mi t1_j786bu0 wrote
Around 2M Edit: the number I gave is completely delusional. Sorry
HeyLittleTrain t1_j7avkil wrote
Your answer seems substantially different than the others.
NapkinsOnMyAnkle t1_j9jtolb wrote
I've trained 100m CNNs on my laptop 3070 6gb. So...
Viewing a single comment thread. View all comments