Optimizers
There are many optimizers:
- Gradient Descent
- Stochastic Gradient Descent (SGD)
- Mini Batch SGD
- Stochastic Gradient Descent with Momentum
- Nesterov Accelerated Gradient (NAG)
- AdaGrad
- RMSProp
- AdaDelta
- Adam
- AdaMax
- NADAM
- AMSGrad
There are many optimizers: