Optimization

[Paper Exploration] Adam: A Method for Stochastic Optimization

From optimization, to convex optimization, to first order optimization, to gradient descent, to accelerated gradient descent, to AdaGrad, to Adam.