Submitted by Vegetable-Skill-9700 t3_121a8p4 in MachineLearning
Yardanico t1_jdls342 wrote
Reply to comment by wojtek15 in [D] Do we really need 100B+ parameters in a large language model? by Vegetable-Skill-9700
Yeah, I think there's a lot of overhyping going around "running ChatGPT-grade language models on consumer hardware". They can "follow" instructions they same way as ChatGPT, but obviously those models know far, far less than the ClosedAI models do, and of course they'll hallucinate much more.
Although it's not an entirely bad thing, at least the community will innovate more so we might get something interesting in the future from this "push" :)
fiftyfourseventeen t1_jdnhbn0 wrote
OpenAI is also doing a lot of tricks behind the scenes, so it's not really fair to just type two things into both, because they are getting nowhere near the same prompt. Llama is promising but it just needs to be properly instruction tuned
devl82 t1_jdmf6b3 wrote
no they is no overhype, you just don't understand what Alpaca is trying to do & I am sure others will also reply similar
Viewing a single comment thread. View all comments