File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: Bagging in overparameterized learning: Risk characterization and risk monotonization

TitleBagging in overparameterized learning: Risk characterization and risk monotonization
Authors
Keywordsdivide-and-conquer
proportional asymptotics
ridge regression
subagging
Issue Date2023
Citation
Journal of Machine Learning Research, 2023, v. 24, article no. 319 How to Cite?
AbstractBagging is a commonly used ensemble technique in statistics and machine learning to improve the performance of prediction procedures. In this paper, we study the prediction risk of variants of bagged predictors under the proportional asymptotics regime, in which the ratio of the number of features to the number of observations converges to a constant. Specifically, we propose a general strategy to analyze the prediction risk under squared error loss of bagged predictors using classical results on simple random sampling. Specializing the strategy, we derive the exact asymptotic risk of the bagged ridge and ridgeless predictors with an arbitrary number of bags under a well-specified linear model with arbitrary feature covariance matrices and signal vectors. Furthermore, we prescribe a generic cross-validation procedure to select the optimal subsample size for bagging and discuss its utility to eliminate the non-monotonic behavior of the limiting risk in the sample size (i.e., double or multiple descents). In demonstrating the proposed procedure for bagged ridge and ridgeless predictors, we thoroughly investigate the oracle properties of the optimal subsample size and provide an in-depth comparison between different bagging variants.
Persistent Identifierhttp://hdl.handle.net/10722/365524
ISSN
2023 Impact Factor: 4.3
2023 SCImago Journal Rankings: 2.796

 

DC FieldValueLanguage
dc.contributor.authorPatil, Pratik-
dc.contributor.authorDu, Jin Hong-
dc.contributor.authorKuchibhotla, Arun Kumar-
dc.date.accessioned2025-11-05T09:41:15Z-
dc.date.available2025-11-05T09:41:15Z-
dc.date.issued2023-
dc.identifier.citationJournal of Machine Learning Research, 2023, v. 24, article no. 319-
dc.identifier.issn1532-4435-
dc.identifier.urihttp://hdl.handle.net/10722/365524-
dc.description.abstractBagging is a commonly used ensemble technique in statistics and machine learning to improve the performance of prediction procedures. In this paper, we study the prediction risk of variants of bagged predictors under the proportional asymptotics regime, in which the ratio of the number of features to the number of observations converges to a constant. Specifically, we propose a general strategy to analyze the prediction risk under squared error loss of bagged predictors using classical results on simple random sampling. Specializing the strategy, we derive the exact asymptotic risk of the bagged ridge and ridgeless predictors with an arbitrary number of bags under a well-specified linear model with arbitrary feature covariance matrices and signal vectors. Furthermore, we prescribe a generic cross-validation procedure to select the optimal subsample size for bagging and discuss its utility to eliminate the non-monotonic behavior of the limiting risk in the sample size (i.e., double or multiple descents). In demonstrating the proposed procedure for bagged ridge and ridgeless predictors, we thoroughly investigate the oracle properties of the optimal subsample size and provide an in-depth comparison between different bagging variants.-
dc.languageeng-
dc.relation.ispartofJournal of Machine Learning Research-
dc.subjectdivide-and-conquer-
dc.subjectproportional asymptotics-
dc.subjectridge regression-
dc.subjectsubagging-
dc.titleBagging in overparameterized learning: Risk characterization and risk monotonization-
dc.typeArticle-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.scopuseid_2-s2.0-85181245424-
dc.identifier.volume24-
dc.identifier.spagearticle no. 319-
dc.identifier.epagearticle no. 319-
dc.identifier.eissn1533-7928-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats