Submitted by singularpanda t3_1060gfk in MachineLearning
f_max t1_j3frqfb wrote
Reply to comment by singularpanda in [D] Will NLP Researchers Lose Our Jobs after ChatGPT? by singularpanda
They have a sequence of models ranging from 6B params up to 175B largest, so you can work on smaller variants if you don’t have gpus. There’s def some papers working on inference efficiency and benchmarking their failure modes if you look around.
Viewing a single comment thread. View all comments