Meta AI introduces LLaMA: A foundational, 65-billion-parameter large language model Submitted by fraktall t3_11b4bim on February 24, 2023 at 10:38 PM in singularity 9 comments 42
Fallen-stars123 t1_j9yufow wrote on February 25, 2023 at 4:01 PM It seems that the new "idea" will be to train a lot more tokens, than just increasing the number of parameters, it seems that we were undertraining the models. I imagine that GPT-4 will see a big jump in the amount of tokens trained. Permalink 2
Viewing a single comment thread. View all comments