On stochastic optimization and the Adam optimizer: Divergence, convergence rates, and acceleration techniques
Dereich & A. Jentzen, Convergence rates for the Adam optimizer, arXiv:2407.21078 (2024), 43 pages. [2] S. Dereich, R. Graeber, & A. Jentzen, Non-convergence of Adam and other adaptive stochastic gradient …