Submitted by Vegetable-Skill-9700 t3_121a8p4 in MachineLearning
gamerx88 t1_jdn1dd3 wrote
Reply to comment by currentscurrents in [D] Do we really need 100B+ parameters in a large language model? by Vegetable-Skill-9700
> In the long run I expect this will flip; computers will get very fast and data will be the limiting factor.
I agree but I think data is already a limiting factor today, with the largest (that is public knowledge) models at 175B. The data used to train these models supposedly already cover a majority of the open internet.
Viewing a single comment thread. View all comments