Links to original paper published on arXiv.org>cs>arXiv:1609.04747 : [1], [2]
Implemented following Gradient Desent Optimization Algorithms from Scratch :
- Vanilla Batch/Stochastic Gradient Descent
- Momentum
- Nesterov Accelarated Gradient
- Adagrad
- Adadelta
- RMS Prop