File Download

There are no files associated with this item.

Supplementary

Conference Paper: Data Pruning via Moving-one-Sample-out

TitleData Pruning via Moving-one-Sample-out
Authors
Issue Date10-Dec-2023
Abstract

In this paper, we propose a novel data-pruning approach called moving-one-sample-out (MoSo), which aims to identify and remove the least informative samples from the training set. The core insight behind MoSo is to determine the importance of each sample by assessing its impact on the optimal empirical risk. This is achieved by measuring the extent to which the empirical risk changes when a particular sample is excluded from the training set. Instead of using the computationally expensive leaving-one-out-retraining procedure, we propose an efficient first-order approximator that only requires gradient information from different training stages. The key idea behind our approximation is that samples with gradients that are consistently aligned with the average gradient of the training set are more informative and should receive higher scores, which could be intuitively understood as follows: if the gradient from a specific sample is consistent with the average gradient vector, it implies that optimizing the network using the sample will yield a similar effect on all remaining samples. Experimental results demonstrate that MoSo effectively mitigates severe performance degradation at high pruning ratios and achieves satisfactory performance across various settings.


Persistent Identifierhttp://hdl.handle.net/10722/339458

 

DC FieldValueLanguage
dc.contributor.authorTan, Haoru-
dc.contributor.authorWu, Sitong-
dc.contributor.authorDu, Fei-
dc.contributor.authorChen, Yukang-
dc.contributor.authorWang, Zhibin-
dc.contributor.authorWang, Fan-
dc.contributor.authorQi, Xiaojuan-
dc.date.accessioned2024-03-11T10:36:47Z-
dc.date.available2024-03-11T10:36:47Z-
dc.date.issued2023-12-10-
dc.identifier.urihttp://hdl.handle.net/10722/339458-
dc.description.abstract<p>In this paper, we propose a novel data-pruning approach called moving-one-sample-out (MoSo), which aims to identify and remove the least informative samples from the training set. The core insight behind MoSo is to determine the importance of each sample by assessing its impact on the optimal empirical risk. This is achieved by measuring the extent to which the empirical risk changes when a particular sample is excluded from the training set. Instead of using the computationally expensive leaving-one-out-retraining procedure, we propose an efficient first-order approximator that only requires gradient information from different training stages. The key idea behind our approximation is that samples with gradients that are consistently aligned with the average gradient of the training set are more informative and should receive higher scores, which could be intuitively understood as follows: if the gradient from a specific sample is consistent with the average gradient vector, it implies that optimizing the network using the sample will yield a similar effect on all remaining samples. Experimental results demonstrate that MoSo effectively mitigates severe performance degradation at high pruning ratios and achieves satisfactory performance across various settings.</p>-
dc.languageeng-
dc.relation.ispartofNeural Information Processing Systems 2023 (10/12/2023-16/12/2023, , , New Orleans)-
dc.titleData Pruning via Moving-one-Sample-out-
dc.typeConference_Paper-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats