Submitted by Available_Lion_652 t3_10xu09v in MachineLearning
Hey guys I have an older PC(5 years) with an i7 7700k processor. I want to buy an Nvidia RTX 3090 for training large language models. I can t find any benchmark for CPU bottleneck when training, let s say an GPT 2 large model. Has anyone have any experience with this set-up similar set-up ?
IntelArtiGen t1_j7uce7z wrote
The CPU bottleneck depends on the model and the training process. If you remove all /most of the preprocessing done on CPU it could be fine. I think transformers don't usually bottleneck on CPU but i7 7700k is quite old.