Submitted by MichelMED10 t3_ysah21 in MachineLearning
Meddhouib10 t1_iw1zmds wrote
Reply to comment by Pretend-Economics758 in [D] Timm's implementation of stochastic depth by MichelMED10
But there is no normalization in dropout right ?
ThomasBudd93 t1_iw228pb wrote
Yes there is. During training with dropout probability p you rescale you inputs with a factor 1/(1-p).
https://pytorch.org/docs/stable/generated/torch.nn.Dropout.html
Viewing a single comment thread. View all comments