File Download

There are no files associated with this item.

Supplementary

Conference Paper: How Does Semi-supervised learning with Pseudo-labelers Work? A Case Study

TitleHow Does Semi-supervised learning with Pseudo-labelers Work? A Case Study
Authors
Issue Date5-May-2023
Abstract

Semi-supervised learning is a popular machine learning paradigm that utilizes a large amount of unlabeled data as well as a small amount of labeled data to facilitate learning tasks. While semi-supervised learning has achieved great success in training neural networks, its theoretical understanding remains largely open. In this paper, we aim to theoretically understand a semi-supervised learning approach based on pre-training and linear probing. In particular, the semi-supervised learning approach we consider first trains a two-layer neural network based on the unlabeled data with the help of pseudo-labelers. Then it linearly probes the pre-trained network on a small amount of labeled data. We prove that, under a certain toy data generation model and two-layer convolutional neural network, the semisupervised learning approach can achieve nearly zero test loss, while a neural network directly trained by supervised learning on the same amount of labeled data can only achieve constant test loss. Through this case study, we demonstrate a separation between semi-supervised learning and supervised learning in terms of test loss provided the same amount of labeled data.


Persistent Identifierhttp://hdl.handle.net/10722/338363

 

DC FieldValueLanguage
dc.contributor.authorKou, Yiwen-
dc.contributor.authorChen, Zixiang-
dc.contributor.authorCao, Yuan-
dc.contributor.authorGu, Quanquan -
dc.date.accessioned2024-03-11T10:28:18Z-
dc.date.available2024-03-11T10:28:18Z-
dc.date.issued2023-05-05-
dc.identifier.urihttp://hdl.handle.net/10722/338363-
dc.description.abstract<p>Semi-supervised learning is a popular machine learning paradigm that utilizes a large amount of unlabeled data as well as a small amount of labeled data to facilitate learning tasks. While semi-supervised learning has achieved great success in training neural networks, its theoretical understanding remains largely open. In this paper, we aim to theoretically understand a semi-supervised learning approach based on pre-training and linear probing. In particular, the semi-supervised learning approach we consider first trains a two-layer neural network based on the unlabeled data with the help of pseudo-labelers. Then it linearly probes the pre-trained network on a small amount of labeled data. We prove that, under a certain toy data generation model and two-layer convolutional neural network, the semisupervised learning approach can achieve nearly zero test loss, while a neural network directly trained by supervised learning on the same amount of labeled data can only achieve constant test loss. Through this case study, we demonstrate a separation between semi-supervised learning and supervised learning in terms of test loss provided the same amount of labeled data.<br></p>-
dc.languageeng-
dc.relation.ispartofThe 11th International Conference on Learning Representations (ICLR 2023) (01/05/2023-05/05/2023, Kigali, Rwanda)-
dc.titleHow Does Semi-supervised learning with Pseudo-labelers Work? A Case Study-
dc.typeConference_Paper-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats