Viewing a single comment thread. View all comments

AiChip t1_j74ku5a wrote

Wow! This is huge! 1B parameters model beating 175 B parameters model…

43

Lengador t1_j74ro7q wrote

That's the number in the headline, but if you look at the tables you can see their 223M parameter model beats the 175B parameter model significantly as well. That's 0.1% the size! Absolutely insane.

53

HeyLittleTrain t1_j77w36w wrote

At what size could I run a model on a decent gaming PC?

8

Lengador t1_j78ovy2 wrote

You can (just) run a 1B parameter model on a good gaming rig.

9

i2mi t1_j786bu0 wrote

Around 2M Edit: the number I gave is completely delusional. Sorry

−4

JClub t1_jabyi76 wrote

GPT was never trained with image data, why is this a fair comparison? The UnifiedQA model is from 2022, so it doesn't seem fair either. Why don't we have some comparisons with other SOTA multimodal models? Such as OFA or UniT

1