Submitted by MichelMED10 t3_ysah21 in MachineLearning
Pretend-Economics758 t1_iw0n3l8 wrote
I guess it’s due to normalisation idea similar to using dropout to reduce overfitting?
Meddhouib10 t1_iw1zmds wrote
But there is no normalization in dropout right ?
ThomasBudd93 t1_iw228pb wrote
Yes there is. During training with dropout probability p you rescale you inputs with a factor 1/(1-p).
https://pytorch.org/docs/stable/generated/torch.nn.Dropout.html
Viewing a single comment thread. View all comments