File Download
There are no files associated with this item.
Supplementary
-
Citations:
- Scopus: 0
- Appears in Collections:
Conference Paper: Fine-Grained analysis of stability and generalization for stochastic gradientdescent
Title | Fine-Grained analysis of stability and generalization for stochastic gradientdescent |
---|---|
Authors | |
Issue Date | 2020 |
Citation | 37th International Conference on Machine Learning, ICML 2020, 2020, v. PartF168147-8, p. 5765-5775 How to Cite? |
Abstract | Recently there are a considerable amount of work devoted to the study of the algorithmic stability and generalization for stochastic gradient descent (SGD). However, the existing stability analysis requires to impose restrictive assumptions on the boundedness of gradients, smoothness and con_vexity of loss functions. In this paper, we provide a fine-grained analysis of stability and general_ization for SGD by substantially relaxing these assumptions. Firstly, we establish stability and generalization for SGD by removing the existing bounded gradient assumptions. The key idea is the introduction of a new stability measure called on-average model stability, for which we develop novel bounds controlled by the risks of SGD iter_ates. This yields generalization bounds depend_ing on the behavior of the best model, and leads to the first-ever-known fast bounds in the low_noise setting using stability approach. Secondly, the smoothness assumption is relaxed by con_sidering loss functions with Holder continuous (sub)gradients for which we show that optimal bounds are still achieved by balancing computa_tion and stability. To our best knowledge, this gives the first-ever-known stability and generaliza_tion bounds for SGD with non-smooth loss func_tions (e.g., hinge loss). Finally, we study learning problems with (strongly) convex objectives but non-convex loss functions. |
Persistent Identifier | http://hdl.handle.net/10722/329705 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Lei, Yunwen | - |
dc.contributor.author | Ying, Yiming | - |
dc.date.accessioned | 2023-08-09T03:34:44Z | - |
dc.date.available | 2023-08-09T03:34:44Z | - |
dc.date.issued | 2020 | - |
dc.identifier.citation | 37th International Conference on Machine Learning, ICML 2020, 2020, v. PartF168147-8, p. 5765-5775 | - |
dc.identifier.uri | http://hdl.handle.net/10722/329705 | - |
dc.description.abstract | Recently there are a considerable amount of work devoted to the study of the algorithmic stability and generalization for stochastic gradient descent (SGD). However, the existing stability analysis requires to impose restrictive assumptions on the boundedness of gradients, smoothness and con_vexity of loss functions. In this paper, we provide a fine-grained analysis of stability and general_ization for SGD by substantially relaxing these assumptions. Firstly, we establish stability and generalization for SGD by removing the existing bounded gradient assumptions. The key idea is the introduction of a new stability measure called on-average model stability, for which we develop novel bounds controlled by the risks of SGD iter_ates. This yields generalization bounds depend_ing on the behavior of the best model, and leads to the first-ever-known fast bounds in the low_noise setting using stability approach. Secondly, the smoothness assumption is relaxed by con_sidering loss functions with Holder continuous (sub)gradients for which we show that optimal bounds are still achieved by balancing computa_tion and stability. To our best knowledge, this gives the first-ever-known stability and generaliza_tion bounds for SGD with non-smooth loss func_tions (e.g., hinge loss). Finally, we study learning problems with (strongly) convex objectives but non-convex loss functions. | - |
dc.language | eng | - |
dc.relation.ispartof | 37th International Conference on Machine Learning, ICML 2020 | - |
dc.title | Fine-Grained analysis of stability and generalization for stochastic gradientdescent | - |
dc.type | Conference_Paper | - |
dc.description.nature | link_to_subscribed_fulltext | - |
dc.identifier.scopus | eid_2-s2.0-85105548031 | - |
dc.identifier.volume | PartF168147-8 | - |
dc.identifier.spage | 5765 | - |
dc.identifier.epage | 5775 | - |