Submitted by Vegetable-Skill-9700 t3_121a8p4 in MachineLearning
PilotThen t1_jdpnoul wrote
Reply to comment by ganzzahl in [D] Do we really need 100B+ parameters in a large language model? by Vegetable-Skill-9700
I didn't find a paper but I think that is sort of what EleutherAI was doing with their pythia models.
You'll find the models on huggingface and I'd say that they are also interesting from an opensource perspective because of their license (apache-2.0)
(Also open-assistent seems to be building on top of them.)
Viewing a single comment thread. View all comments