Viewing a single comment thread. View all comments

Acceptable-Cress-374 t1_ixbzdfe wrote

Would this mean that it could become feasible to run gpt-neox inference on a 3090/4090 w/ 24 GB VRAM? That would be huge!

8