File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1162/neco_a_01588
- Scopus: eid_2-s2.0-85161145576
- PMID: 37187168
- WOS: WOS:001125378300004
- Find via
Supplementary
- Citations:
- Appears in Collections:
Article: Optimization and Learning With Randomly Compressed Gradient Updates
Title | Optimization and Learning With Randomly Compressed Gradient Updates |
---|---|
Authors | |
Issue Date | 2023 |
Citation | Neural Computation, 2023, v. 35, n. 7, p. 1234-1287 How to Cite? |
Abstract | Gradient descent methods are simple and efficient optimization algorithms with widespread applications. To handle high-dimensional prob-lems, we study compressed stochastic gradient descent (SGD) with low-dimensional gradient updates. We provide a detailed analysis in terms of both optimization rates and generalization rates. To this end, we develop uniform stability bounds for CompSGD for both smooth and nonsmooth problems, based on which we develop almost optimal population risk bounds. Then we extend our analysis to two variants of SGD: batch and mini-batch gradient descent. Furthermore, we show that these variants achieve almost optimal rates compared to their high-dimensional gradient setting. Thus, our results provide a way to reduce the dimension of gradient updates without affecting the convergence rate in the generalization analysis. Moreover, we show that the same result also holds in the differentially private setting, which allows us to reduce the dimension of added noise with “almost free” cost. |
Persistent Identifier | http://hdl.handle.net/10722/329976 |
ISSN | 2023 Impact Factor: 2.7 2023 SCImago Journal Rankings: 0.948 |
ISI Accession Number ID |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Huang, Zhanliang | - |
dc.contributor.author | Lei, Yunwen | - |
dc.contributor.author | Kabán, Ata | - |
dc.date.accessioned | 2023-08-09T03:36:55Z | - |
dc.date.available | 2023-08-09T03:36:55Z | - |
dc.date.issued | 2023 | - |
dc.identifier.citation | Neural Computation, 2023, v. 35, n. 7, p. 1234-1287 | - |
dc.identifier.issn | 0899-7667 | - |
dc.identifier.uri | http://hdl.handle.net/10722/329976 | - |
dc.description.abstract | Gradient descent methods are simple and efficient optimization algorithms with widespread applications. To handle high-dimensional prob-lems, we study compressed stochastic gradient descent (SGD) with low-dimensional gradient updates. We provide a detailed analysis in terms of both optimization rates and generalization rates. To this end, we develop uniform stability bounds for CompSGD for both smooth and nonsmooth problems, based on which we develop almost optimal population risk bounds. Then we extend our analysis to two variants of SGD: batch and mini-batch gradient descent. Furthermore, we show that these variants achieve almost optimal rates compared to their high-dimensional gradient setting. Thus, our results provide a way to reduce the dimension of gradient updates without affecting the convergence rate in the generalization analysis. Moreover, we show that the same result also holds in the differentially private setting, which allows us to reduce the dimension of added noise with “almost free” cost. | - |
dc.language | eng | - |
dc.relation.ispartof | Neural Computation | - |
dc.title | Optimization and Learning With Randomly Compressed Gradient Updates | - |
dc.type | Article | - |
dc.description.nature | link_to_subscribed_fulltext | - |
dc.identifier.doi | 10.1162/neco_a_01588 | - |
dc.identifier.pmid | 37187168 | - |
dc.identifier.scopus | eid_2-s2.0-85161145576 | - |
dc.identifier.volume | 35 | - |
dc.identifier.issue | 7 | - |
dc.identifier.spage | 1234 | - |
dc.identifier.epage | 1287 | - |
dc.identifier.eissn | 1530-888X | - |
dc.identifier.isi | WOS:001125378300004 | - |