File Download
  Links for fulltext
     (May Require Subscription)
Supplementary

Article: Learning Rates for Stochastic Gradient Descent with Nonconvex Objectives

TitleLearning Rates for Stochastic Gradient Descent with Nonconvex Objectives
Authors
Keywordsearly stopping
learning rates
nonconvex optimization
Stochastic gradient descent
Issue Date21-Mar-2021
PublisherInstitute of Electrical and Electronics Engineers
Citation
IEEE Transactions on Pattern Analysis and Machine Intelligence, 2021, v. 43, n. 12, p. 4505-4511 How to Cite?
Abstract

Stochastic gradient descent (SGD) has become the method of choice for training highly complex and nonconvex models since it can not only recover good solutions to minimize training errors but also generalize well. Computational and statistical properties are separately studied to understand the behavior of SGD in the literature. However, there is a lacking study to jointly consider the computational and statistical properties in a nonconvex learning setting. In this paper, we develop novel learning rates of SGD for nonconvex learning by presenting high-probability bounds for both computational and statistical errors. We show that the complexity of SGD iterates grows in a controllable manner with respect to the iteration number, which sheds insights on how an implicit regularization can be achieved by tuning the number of passes to balance the computational and statistical errors. As a byproduct, we also slightly refine the existing studies on the uniform convergence of gradients by showing its connection to Rademacher chaos complexities.


Persistent Identifierhttp://hdl.handle.net/10722/329132
ISSN
2021 Impact Factor: 24.314
2020 SCImago Journal Rankings: 3.811

 

DC FieldValueLanguage
dc.contributor.authorLei, Yunwen-
dc.contributor.authorTang, Ke -
dc.date.accessioned2023-08-05T07:55:32Z-
dc.date.available2023-08-05T07:55:32Z-
dc.date.issued2021-03-21-
dc.identifier.citationIEEE Transactions on Pattern Analysis and Machine Intelligence, 2021, v. 43, n. 12, p. 4505-4511-
dc.identifier.issn0162-8828-
dc.identifier.urihttp://hdl.handle.net/10722/329132-
dc.description.abstract<p>Stochastic gradient descent (SGD) has become the method of choice for training highly complex and nonconvex models since it can not only recover good solutions to minimize training errors but also generalize well. Computational and statistical properties are separately studied to understand the behavior of SGD in the literature. However, there is a lacking study to jointly consider the computational and statistical properties in a nonconvex learning setting. In this paper, we develop novel learning rates of SGD for nonconvex learning by presenting high-probability bounds for both computational and statistical errors. We show that the complexity of SGD iterates grows in a controllable manner with respect to the iteration number, which sheds insights on how an implicit regularization can be achieved by tuning the number of passes to balance the computational and statistical errors. As a byproduct, we also slightly refine the existing studies on the uniform convergence of gradients by showing its connection to Rademacher chaos complexities.<br></p>-
dc.languageeng-
dc.publisherInstitute of Electrical and Electronics Engineers-
dc.relation.ispartofIEEE Transactions on Pattern Analysis and Machine Intelligence-
dc.rightsThis work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.-
dc.subjectearly stopping-
dc.subjectlearning rates-
dc.subjectnonconvex optimization-
dc.subjectStochastic gradient descent-
dc.titleLearning Rates for Stochastic Gradient Descent with Nonconvex Objectives-
dc.typeArticle-
dc.description.naturepublished_or_final_version-
dc.identifier.doi10.1109/TPAMI.2021.3068154-
dc.identifier.scopuseid_2-s2.0-85103253583-
dc.identifier.volume43-
dc.identifier.issue12-
dc.identifier.spage4505-
dc.identifier.epage4511-
dc.identifier.eissn1939-3539-
dc.identifier.issnl0162-8828-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats