Neural Network architectures

CNN基础(2)

a mathematical model of neuron


Activation Functions

  • Sigmoid

    σ(x)=1/(1+ex)

    CNN基础(2)

    • Sigmoids saturate and kill gradients
    • Sigmoid outputs are not zero-centered
  • tanh

    tanh(x)=2σ(2x)1

    CNN基础(2)

    • in practice the tanh non-linearity is always preferred to the sigmoid nonlinearity
  • ReLU(Rectified Linear Unit )

    f(x)=max(0,x)

    CNN基础(2)CNN基础(2)

    • (+) It was found to greatly accelerate
    • (+) ReLU can be implemented by simply thresholding a matrix of activations at zero
    • (-)Unfortunately, ReLU units can be fragile during training and can “die”
  • Leaky ReLU

    f(x)=

相关文章:

  • 2021-06-08
  • 2022-01-02
  • 2021-07-26
  • 2021-12-09
  • 2021-06-18
  • 2022-01-26
  • 2022-12-23
  • 2022-12-23
猜你喜欢
  • 2021-06-28
  • 2021-11-25
  • 2022-01-16
  • 2022-01-06
  • 2021-05-29
  • 2021-06-28
相关资源
相似解决方案