If we set the initial \(\Theta\) be the same, the units in next layer with the same \(x_i\) will get the same result, then all units in the same layer will get the same output. At last, the cost function will also get same cost, so we will update the \(\Theta\) with same step.
seems \(\delta^{(l)}_{i}\) means the cost of \(i_{th}\) unit in the \(l_{th}\) layer
Security Info Adversary Learning TCP/IP Pattern Recognition Python OS Deep Learning Machine Learning
阿毛
Copyright © 2016 Powered by MWeb, Theme used GitHub CSS. Modified by Yinzo.