Stay Hungry,Stay Foolish!

Various Optimization Algorithms For Training Neural Network

from

https://towardsdatascience.com/optimizers-for-training-neural-network-59450d71caf6

 

 

Optimizers help to get results faster

Gradient Descent

Stochastic Gradient Descent

Mini-Batch Gradient Descent

Momentum

Nesterov Accelerated Gradient

 

NAG vs momentum at local minima

Adagrad

 

A derivative of loss function for given parameters at a given time t.

 

Update parameters for given input i and at time/iteration t

AdaDelta

 

Update the parameters

Adam

 

First and second order of momentum

Update the parameters

Comparison between various optimizers

 

Comparison 1

 

comparison 2

Conclusions

 

posted @ 2021-04-11 10:15  lightsong  阅读(69)  评论(0编辑  收藏  举报
Life Is Short, We Need Ship To Travel