File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: Per-example gradient regularization improves learning signals from noisy data

TitlePer-example gradient regularization improves learning signals from noisy data
Authors
KeywordsGradient regularization
Noise perturbations
Variance control
Issue Date1-Mar-2025
PublisherSpringer
Citation
Machine Learning, 2025, v. 114, n. 3 How to Cite?
AbstractGradient regularization, as described in Barrett and Dherin (in: International conference on learning representations, 2021), is a highly effective technique for promoting flat minima during gradient descent. Empirical evidence suggests that this regularization technique can significantly enhance the robustness of deep learning models against noisy perturbations, while also reducing test error. In this paper, we explore the per-example gradient regularization (PEGR) and present a theoretical analysis that demonstrates its effectiveness in improving both test error and robustness against noise perturbations. Specifically, we adopt a signal-noise data model from Cao et al. (Adv Neural Inf Process Syst 35:25237–25250, 2022) and show that PEGR can learn signals effectively while suppressing noise memorization. In contrast, standard gradient descent struggles to distinguish the signal from the noise, leading to suboptimal generalization performance. Our analysis reveals that PEGR penalizes the variance of pattern learning, thus effectively suppressing the memorization of noises from the training data. These findings underscore the importance of variance control in deep learning training and offer useful insights for developing more effective training approaches.
Persistent Identifierhttp://hdl.handle.net/10722/358381
ISSN
2023 Impact Factor: 4.3
2023 SCImago Journal Rankings: 1.720

 

DC FieldValueLanguage
dc.contributor.authorMeng, Xuran-
dc.contributor.authorCao, Yuan-
dc.contributor.authorZou, Difan-
dc.date.accessioned2025-08-07T00:31:53Z-
dc.date.available2025-08-07T00:31:53Z-
dc.date.issued2025-03-01-
dc.identifier.citationMachine Learning, 2025, v. 114, n. 3-
dc.identifier.issn0885-6125-
dc.identifier.urihttp://hdl.handle.net/10722/358381-
dc.description.abstractGradient regularization, as described in Barrett and Dherin (in: International conference on learning representations, 2021), is a highly effective technique for promoting flat minima during gradient descent. Empirical evidence suggests that this regularization technique can significantly enhance the robustness of deep learning models against noisy perturbations, while also reducing test error. In this paper, we explore the per-example gradient regularization (PEGR) and present a theoretical analysis that demonstrates its effectiveness in improving both test error and robustness against noise perturbations. Specifically, we adopt a signal-noise data model from Cao et al. (Adv Neural Inf Process Syst 35:25237–25250, 2022) and show that PEGR can learn signals effectively while suppressing noise memorization. In contrast, standard gradient descent struggles to distinguish the signal from the noise, leading to suboptimal generalization performance. Our analysis reveals that PEGR penalizes the variance of pattern learning, thus effectively suppressing the memorization of noises from the training data. These findings underscore the importance of variance control in deep learning training and offer useful insights for developing more effective training approaches.-
dc.languageeng-
dc.publisherSpringer-
dc.relation.ispartofMachine Learning-
dc.rightsThis work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.-
dc.subjectGradient regularization-
dc.subjectNoise perturbations-
dc.subjectVariance control-
dc.titlePer-example gradient regularization improves learning signals from noisy data -
dc.typeArticle-
dc.identifier.doi10.1007/s10994-024-06661-5-
dc.identifier.scopuseid_2-s2.0-85218201264-
dc.identifier.volume114-
dc.identifier.issue3-
dc.identifier.eissn1573-0565-
dc.identifier.issnl0885-6125-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats