- Journal Home
- Volume 18 - 2025
- Volume 17 - 2024
- Volume 16 - 2023
- Volume 15 - 2022
- Volume 14 - 2021
- Volume 13 - 2020
- Volume 12 - 2019
- Volume 11 - 2018
- Volume 10 - 2017
- Volume 9 - 2016
- Volume 8 - 2015
- Volume 7 - 2014
- Volume 6 - 2013
- Volume 5 - 2012
- Volume 4 - 2011
- Volume 3 - 2010
- Volume 2 - 2009
- Volume 1 - 2008
Numer. Math. Theor. Meth. Appl., 16 (2023), pp. 914-930.
Published online: 2023-11
Cited by
- BibTex
- RIS
- TXT
In this paper, we first reinvestigate the convergence of the vanilla SGD method in the sense of $L^2$ under more general learning rates conditions and a more general convex assumption, which relieves the conditions on learning rates and does not need the problem to be strongly convex. Then, by taking advantage of the Lyapunov function technique, we present the convergence of the momentum SGD and Nesterov accelerated SGD methods for the convex and non-convex problem under $L$-smooth assumption that extends the bounded gradient limitation to a certain extent. The convergence of time averaged SGD was also analyzed.
}, issn = {2079-7338}, doi = {https://doi.org/10.4208/nmtma.OA-2022-0179}, url = {http://global-sci.org/intro/article_detail/nmtma/22116.html} }In this paper, we first reinvestigate the convergence of the vanilla SGD method in the sense of $L^2$ under more general learning rates conditions and a more general convex assumption, which relieves the conditions on learning rates and does not need the problem to be strongly convex. Then, by taking advantage of the Lyapunov function technique, we present the convergence of the momentum SGD and Nesterov accelerated SGD methods for the convex and non-convex problem under $L$-smooth assumption that extends the bounded gradient limitation to a certain extent. The convergence of time averaged SGD was also analyzed.