Submitted by thanderrine t3_zc0kco in MachineLearning
Hey people,
So. I wanted to find out if there is a way to determine when to early stop a training job.
See let's say I'm running the job for 100 epochs, the graph between training and validation accuracy and training and validation loss flattens at about 91% leading to drumroll over fitting! (Obviously).
Now apart from dropout layer, I'm using early stopping. But the issue is, I'm kind of concerned that it finds a local minimum and stops execution.
I'm using validation loss BTW for early stopping.
ppg_dork t1_iyue984 wrote
I think the answer is going to depend heavily on your domain. For the types of datasets I typically use, I will keep the training going for a while, drop the learning rate once or twice if the validation loss keeps going down, and eventually stop when overfitting begins or loss plateaus.
However, I typically work on (relatively) simple CV problems. I've heard from colleagues that they sometimes train well beyond validation loss beginning to increase as the loss will eventually drop back down. However, this seems more common with RL or generative models. I'll reiterate that I'm regurgitating what others have mentioned.
If your goal ISN'T to shoot for SOTA, I would look at the "best" examples of applied literature in a relevant field. I've often found that those papers provide recipes that transfer well to similar problems. Whereas the SOTA-setting papers tend to be quite specialized to ImageNet/CIFAR-whatever/etc.