Submitted by MLNoober t3_xuogm3 in MachineLearning
jms4607 t1_iqxuph2 wrote
Reply to comment by bushrod in [D] Why restrict to using a linear function to represent neurons? by MLNoober
Generalization out of distribution might be the biggest thing holding back ML rn. It’s worth thinking about whether the priors we encode in nns now are to blame. A large mlp is required just to approximate a single neuron. Maybe the unit additive nonlinearity we are using now is too simple. I’m sure there is a sweet spot between complex interactions/few neurons and simple interactions/many neurons.
Viewing a single comment thread. View all comments