Submitted by MLNoober t3_xuogm3 in MachineLearning
MrFlufypants t1_iqx6bhc wrote
Reply to comment by ZestyData in [D] Why restrict to using a linear function to represent neurons? by MLNoober
The activation functions are key. A linear combination of linear combinations is probably equal to a linear combination, so 10 layers would equate to a single layer, which is only capable of so much. The activation functions destroy the linearity though and are the key ingredient there
Viewing a single comment thread. View all comments