Submitted by lambolifeofficial t3_zzn35o in MachineLearning
Disastrous_Elk_6375 t1_j2de4o2 wrote
Reply to comment by ThatInternetGuy in An Open-Source Version of ChatGPT is Coming [News] by lambolifeofficial
Can the 4090 pool their VRAM? I always thought that LLMs need GPUs from the A/V series so that they can pool memory. Am I wrong in thinking that?
zaptrem t1_j2e2lvb wrote
You can do pipeline parallelism via FairScale and HF Accelerate on any identical (and sometimes non identical) GPUs.
ThatInternetGuy t1_j2deqmr wrote
Need to deploy the inference model with Colossal AI.
Viewing a single comment thread. View all comments