File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Sharper generalization bounds for pairwise learning

TitleSharper generalization bounds for pairwise learning
Authors
Issue Date2020
Citation
Advances in Neural Information Processing Systems, 2020, v. 2020-December How to Cite?
AbstractPairwise learning refers to learning tasks with loss functions depending on a pair of training examples, which includes ranking and metric learning as specific examples. Recently, there has been an increasing amount of attention on the generalization analysis of pairwise learning to understand its practical behavior. However, the existing stability analysis provides suboptimal high-probability generalization bounds. In this paper, we provide a refined stability analysis by developing generalization bounds which can be vn-times faster than the existing results, where n is the sample size. This implies excess risk bounds of the order O(n-1/2) (up to a logarithmic factor) for both regularized risk minimization and stochastic gradient descent. We also introduce a new on-average stability measure to develop optimistic bounds in a low noise setting. We apply our results to ranking and metric learning, and clearly show the advantage of our generalization bounds over the existing analysis.
Persistent Identifierhttp://hdl.handle.net/10722/329691
ISSN
2020 SCImago Journal Rankings: 1.399

 

DC FieldValueLanguage
dc.contributor.authorLei, Yunwen-
dc.contributor.authorLedent, Antoine-
dc.contributor.authorKloft, Marius-
dc.date.accessioned2023-08-09T03:34:38Z-
dc.date.available2023-08-09T03:34:38Z-
dc.date.issued2020-
dc.identifier.citationAdvances in Neural Information Processing Systems, 2020, v. 2020-December-
dc.identifier.issn1049-5258-
dc.identifier.urihttp://hdl.handle.net/10722/329691-
dc.description.abstractPairwise learning refers to learning tasks with loss functions depending on a pair of training examples, which includes ranking and metric learning as specific examples. Recently, there has been an increasing amount of attention on the generalization analysis of pairwise learning to understand its practical behavior. However, the existing stability analysis provides suboptimal high-probability generalization bounds. In this paper, we provide a refined stability analysis by developing generalization bounds which can be vn-times faster than the existing results, where n is the sample size. This implies excess risk bounds of the order O(n-1/2) (up to a logarithmic factor) for both regularized risk minimization and stochastic gradient descent. We also introduce a new on-average stability measure to develop optimistic bounds in a low noise setting. We apply our results to ranking and metric learning, and clearly show the advantage of our generalization bounds over the existing analysis.-
dc.languageeng-
dc.relation.ispartofAdvances in Neural Information Processing Systems-
dc.titleSharper generalization bounds for pairwise learning-
dc.typeConference_Paper-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.scopuseid_2-s2.0-85102123715-
dc.identifier.volume2020-December-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats