Submitted by thanderrine t3_zc0kco in MachineLearning
Hey people,
So. I wanted to find out if there is a way to determine when to early stop a training job.
See let's say I'm running the job for 100 epochs, the graph between training and validation accuracy and training and validation loss flattens at about 91% leading to drumroll over fitting! (Obviously).
Now apart from dropout layer, I'm using early stopping. But the issue is, I'm kind of concerned that it finds a local minimum and stops execution.
I'm using validation loss BTW for early stopping.
_Arsenie_Boca_ t1_iyuu6le wrote
I usually prefer checkpointing over early stopping, i.e. you always save a checkpoint when you get a better validation score. Loss is typically a good indicator, but if you have a more specific measure that you are aiming for(like downstream metrics), you should use that.