Submitted by yazriel0 t3_10nhbfl in MachineLearning
londons_explorer t1_j6al3tb wrote
Reply to comment by mocny-chlapik in [N] OpenAI has 1000s of contractors to fine-tune codex by yazriel0
>They were not able to find significant improvements with scaling anymore.
GPT-3 has a window size of 2048 tokens ChatGPT has a window size of 8192 tokens. The compute cost is superliner, so I suspect the compute required for ChatGPT is a minimum of 10x what GPT-3 used. And GPT-3 cost ~12M USD. (At market rates - I assume they got a deep discount)
So I suspect they did scale compute as much as they could afford.
Viewing a single comment thread. View all comments