File Download
There are no files associated with this item.
Supplementary
-
Citations:
- Appears in Collections:
Conference Paper: Understanding Train-Validation Split in Meta-Learning with Neural Networks
Title | Understanding Train-Validation Split in Meta-Learning with Neural Networks |
---|---|
Authors | |
Issue Date | 5-May-2023 |
Abstract | The goal of meta-learning is to learn a good prior model from a collection of tasks such that the learned prior is able to adapt quickly to new tasks without accessing many data from the new tasks. A common practice in meta-learning is to perform a train-validation split on each task, where the training set is used for adapting the model parameter to that specific task and the validation set is used for learning a prior model that is shared across all tasks. Despite its success and popularity in multitask learning and few-shot learning, the understanding of the train-validation split is still limited, especially when the neural network models are used. In this paper, we study the benefit of train-validation split for classification problems with neural network models trained by gradient descent. We prove that the train-validation split is necessary to learn a good prior model when the noise in the training sample is large, while the train-train method fails. We validate our theory by conducting experiment on both synthetic and real datasets. To the best of our knowledge, this is the first work towards the theoretical understanding of train-validation split in meta-learning with neural networks. |
Persistent Identifier | http://hdl.handle.net/10722/338365 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Zuo, Xinzhe | - |
dc.contributor.author | Chen, Zixiang | - |
dc.contributor.author | Yao, Huaxiu | - |
dc.contributor.author | Cao, Yuan | - |
dc.contributor.author | Gu, Quanquan | - |
dc.date.accessioned | 2024-03-11T10:28:19Z | - |
dc.date.available | 2024-03-11T10:28:19Z | - |
dc.date.issued | 2023-05-05 | - |
dc.identifier.uri | http://hdl.handle.net/10722/338365 | - |
dc.description.abstract | <p>The goal of meta-learning is to learn a good prior model from a collection of tasks such that the learned prior is able to adapt quickly to new tasks without accessing many data from the new tasks. A common practice in meta-learning is to perform a train-validation split on each task, where the training set is used for adapting the model parameter to that specific task and the validation set is used for learning a prior model that is shared across all tasks. Despite its success and popularity in multitask learning and few-shot learning, the understanding of the train-validation split is still limited, especially when the neural network models are used. In this paper, we study the benefit of train-validation split for classification problems with neural network models trained by gradient descent. We prove that the train-validation split is necessary to learn a good prior model when the noise in the training sample is large, while the train-train method fails. We validate our theory by conducting experiment on both synthetic and real datasets. To the best of our knowledge, this is the first work towards the theoretical understanding of train-validation split in meta-learning with neural networks.</p> | - |
dc.language | eng | - |
dc.relation.ispartof | The 11th International Conference on Learning Representations (ICLR 2023) (01/05/2023-05/05/2023, Kigali, Rwanda) | - |
dc.title | Understanding Train-Validation Split in Meta-Learning with Neural Networks | - |
dc.type | Conference_Paper | - |