Submitted by AylaDoesntLikeYou t3_11c5n1g in singularity
visarga t1_ja2u514 wrote
Reply to comment by duffmanhb in Meta unveils a new large language model that can run on a single GPU by AylaDoesntLikeYou
But they documented how to make it by sharing paper, code, dataset and hyper-parameters. So when Stability wants to replicate, it will be 10x cheaper. And they showed a small model can be surprisingly good, that means it is tempting for many to replicate it.
The cost of running inference on GPT-3 was a huge moat that is going away. I expect this year we will be able to run a chatGPT level model on a single GPU, so we get cheap to run, private, open and commercial AI soon. We can use it for ourselves, we can make projects with it.
duffmanhb t1_ja2ugba wrote
I hope so. I'm still waiting for them to accept my invite. But soon as I get it, first thing I'll do is create some llama bots for Reddit and see how effective it is compared to GPT3 posting believable comments. If it's nearly as good, but can be ran locally, it'll completely change the bot game on social media.
Viewing a single comment thread. View all comments