神经网络
Neural Network
Activation Function | 激活函数
Sigmod
神经网络中的激活函数,其作用就是引入非线性。具体的非线性形式,则有多种选择。
$$ S(x) = \frac{1}{1+ e^{-x}} $$
Geoff Hinton covered exactly this topic in his coursera course on neural nets. The problem with sigmoids is that as you reach saturation (values get close to 1 or 0), the gradients vanish. This is detrimental to optimization speed. Softmax doesn’t have this problem, and in fact if you combine softmax with a cross entropy error function the gradients are just (z-y), as they would be for a linear output with least squares error.
其导数