[D] Build a home PC to Run Large GPT Models or use AWS Submitted by [deleted] t3_zrx665 on December 21, 2022 at 6:46 PM in MachineLearning 9 comments 0
LetterRip t1_j164stx wrote on December 21, 2022 at 11:08 PM If slow training is acceptable you can use DeepSpeed with the weights mapped to the NVME drive (DeepSpeed ZeRo Infinity). It will take significantly longer to fine tune, but dramatically lowers the hardware investment. Permalink 1
Viewing a single comment thread. View all comments