Submitted by head_robotics t3_1172jrs in MachineLearning
I've been looking into open source large language models to run locally on my machine.
Seems GPT-J and GPT-Neo are out of reach for me because of RAM / VRAM requirements.
What models would be doable with this hardware?:
CPU: AMD Ryzen 7 3700X 8-Core, 3600 MhzRAM: 32 GB
GPUs:
- NVIDIA GeForce RTX 2070 8GB VRAM
- NVIDIA Tesla M40 24GB VRAM
Disastrous_Elk_6375 t1_j99ry6s wrote
GPT-NeoX should fit in 24GB VRAM with 8bit, for inference.
I managed to run GPT-J 6B on a 3060 w/ 12GB and it takes about 7.2GB of VRAM.