File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: Convergence of Adaptive Stochastic Mirror Descent

TitleConvergence of Adaptive Stochastic Mirror Descent
Authors
KeywordsAdam
convergence analysis
mirror descent
nonconvex stochastic optimization
Issue Date18-Mar-2025
PublisherInstitute of Electrical and Electronics Engineers
Citation
IEEE Transactions on Neural Networks and Learning Systems, 2025, p. 1-12 How to Cite?
AbstractIn this article, we present a family of adaptive stochastic optimization methods, which are associated with mirror maps that are widely used to capture the geometry properties of optimization problems during iteration processes. The wellknown adaptive moment estimation (Adam)-type algorithm falls into the family when the mirror maps take the form of temporal adaptation. In the context of convex objective functions, we show that with proper step sizes and hyperparameters, the average regret can achieve the convergence rate O(T-(1=2)) after T iterations under some standard assumptions. We further improve it to O(T-1 log T) when the objective functions are strongly convex. In the context of smooth objective functions (not necessarily convex), based on properties of the strongly convex differentiable mirror map, our algorithms achieve convergence rates of order O(T-(1=2)) up to a logarithmic term, requiring large or increasing hyperparameters that are coincident with practical usage of Adam-type algorithms. Thus, our work gives explanations for the selection of the hyperparameters in Adamtype algorithms' implementation.
Persistent Identifierhttp://hdl.handle.net/10722/357587
ISSN
2023 Impact Factor: 10.2
2023 SCImago Journal Rankings: 4.170
ISI Accession Number ID

 

DC FieldValueLanguage
dc.contributor.authorHu, Ting-
dc.contributor.authorLiu, Xiaotong-
dc.contributor.authorJi, Kai-
dc.contributor.authorLei, Yunwen-
dc.date.accessioned2025-07-22T03:13:40Z-
dc.date.available2025-07-22T03:13:40Z-
dc.date.issued2025-03-18-
dc.identifier.citationIEEE Transactions on Neural Networks and Learning Systems, 2025, p. 1-12-
dc.identifier.issn2162-237X-
dc.identifier.urihttp://hdl.handle.net/10722/357587-
dc.description.abstractIn this article, we present a family of adaptive stochastic optimization methods, which are associated with mirror maps that are widely used to capture the geometry properties of optimization problems during iteration processes. The wellknown adaptive moment estimation (Adam)-type algorithm falls into the family when the mirror maps take the form of temporal adaptation. In the context of convex objective functions, we show that with proper step sizes and hyperparameters, the average regret can achieve the convergence rate O(T-(1=2)) after T iterations under some standard assumptions. We further improve it to O(T-1 log T) when the objective functions are strongly convex. In the context of smooth objective functions (not necessarily convex), based on properties of the strongly convex differentiable mirror map, our algorithms achieve convergence rates of order O(T-(1=2)) up to a logarithmic term, requiring large or increasing hyperparameters that are coincident with practical usage of Adam-type algorithms. Thus, our work gives explanations for the selection of the hyperparameters in Adamtype algorithms' implementation.-
dc.languageeng-
dc.publisherInstitute of Electrical and Electronics Engineers-
dc.relation.ispartofIEEE Transactions on Neural Networks and Learning Systems-
dc.subjectAdam-
dc.subjectconvergence analysis-
dc.subjectmirror descent-
dc.subjectnonconvex stochastic optimization-
dc.titleConvergence of Adaptive Stochastic Mirror Descent-
dc.typeArticle-
dc.identifier.doi10.1109/TNNLS.2025.3545420-
dc.identifier.scopuseid_2-s2.0-105000426521-
dc.identifier.spage1-
dc.identifier.epage12-
dc.identifier.eissn2162-2388-
dc.identifier.isiWOS:001470663200001-
dc.identifier.issnl2162-237X-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats