Submitted by LightGreenSquash t3_yrsqcz in MachineLearning
LightGreenSquash OP t1_iwi9q1g wrote
Reply to comment by banmeyoucoward in [D] "Grokking" Deep Learning architectures and using them in practice by LightGreenSquash
Yep, that's kind of along the lines I'm thinking as well. The only possible drawback I can see is that for such small datasets even "basic" architectures like MLPs can do well enough and thus you might not be able to see the benefit, say, a ResNet brings.
It's still very much a solid approach though, and I've used it in the past to deepen my knowledge of stuff I already knew, e.g. coding a very basic computational graph framework and then using it to train an MLP on MNIST. It was really cool to see my "hand-made" graph topological sort + fprop/bprop methods written for different functions actually reach 90%+ accuracy.
Viewing a single comment thread. View all comments