Submitted by MyActualUserName99 t3_10mmniu in MachineLearning
MadScientist-1214 t1_j6433qc wrote
At my institute, nobody trained on ImageNet, so I had to figure it out myself too. If you train on architectures like VGG, it does not take long. <2 days on a single A100, with worse GPU max. 5 days. The most important thing is to use SSD, this increases speed by around 2 days. A good learning scheduler is really important. Most researchers ignore the test set, use only validation set. And also important: use mixed precision. You should really tune the training speed, if you need to do a lot of experiments.
Viewing a single comment thread. View all comments