[D] List of Large Language Models to play with. Submitted by sinavski t3_10uh62c on February 5, 2023 at 4:54 PM in MachineLearning 22 comments 86
mrpogiface t1_j7g03gj wrote on February 6, 2023 at 3:09 PM Do we actually know that chatGPT is the full 175B? With codex being 13B and still enormously powerful, and previous instruction tuned models (in the paper) being 6.7B it seems likely that they have it working on a much smaller parameter count Permalink 7
Viewing a single comment thread. View all comments