summary

2.1 mini-batch gradient

the size of batch: m BGD: too long for each iteration

the size of batch:1 SGD: lose speed up(vectorization)

in-between mini-batch


2.2 bias correct

For the beginning of the curve, we use bias correct.( the curve doesn't fit with the former data)

[coursera/ImprovingDL/week2]Optimization algorithms(summary&question)

2.3 exponentially weighted averages

[coursera/ImprovingDL/week2]Optimization algorithms(summary&question)

2.4 gd with momentum

[coursera/ImprovingDL/week2]Optimization algorithms(summary&question)

2.5 RMSprop

[coursera/ImprovingDL/week2]Optimization algorithms(summary&question)

2.6 Adam

momentum+RMSprop+bias correct

[coursera/ImprovingDL/week2]Optimization algorithms(summary&question)


2.7 learning rate decay

[coursera/ImprovingDL/week2]Optimization algorithms(summary&question)

[coursera/ImprovingDL/week2]Optimization algorithms(summary&question)


question:

[coursera/ImprovingDL/week2]Optimization algorithms(summary&question)

[coursera/ImprovingDL/week2]Optimization algorithms(summary&question)

[coursera/ImprovingDL/week2]Optimization algorithms(summary&question)

[coursera/ImprovingDL/week2]Optimization algorithms(summary&question)

[coursera/ImprovingDL/week2]Optimization algorithms(summary&question)

[coursera/ImprovingDL/week2]Optimization algorithms(summary&question)

[coursera/ImprovingDL/week2]Optimization algorithms(summary&question)

[coursera/ImprovingDL/week2]Optimization algorithms(summary&question)

[coursera/ImprovingDL/week2]Optimization algorithms(summary&question)

[coursera/ImprovingDL/week2]Optimization algorithms(summary&question)[coursera/ImprovingDL/week2]Optimization algorithms(summary&question)

[coursera/ImprovingDL/week2]Optimization algorithms(summary&question)




相关文章:

  • 2021-05-19
  • 2021-10-16
  • 2021-11-24
  • 2021-11-11
  • 2021-04-22
  • 2021-05-15
  • 2021-09-29
猜你喜欢
  • 2021-04-13
  • 2022-01-07
  • 2021-04-17
  • 2021-06-23
  • 2021-12-15
  • 2021-12-29
  • 2022-01-15
相关资源
相似解决方案