File Download

There are no files associated with this item.

Conference Paper: The Implicit Bias of Batch Normalization in Linear Models and Two-layer Linear Convolutional Neural Networks

TitleThe Implicit Bias of Batch Normalization in Linear Models and Two-layer Linear Convolutional Neural Networks
Authors
Issue Date15-Jul-2023
Abstract

We study the implicit bias of batch normalization trained by gradient descent. We show that when learning a linear model with batch normalization for binary classification, gradient descent converges to a uniform margin classifier on the training data with an exp(−Ω(log2t)) convergence rate. This distinguishes linear models with batch normalization from those without batch normalization in terms of both the type of implicit bias and the convergence rate. We further extend our result to a class of two-layer, single-filter linear convolutional neural networks, and show that batch normalization has an implicit bias towards a patch-wise uniform margin. Based on two examples, we demonstrate that patch-wise uniform margin classifiers can outperform the maximum margin classifiers in certain learning problems. Our results contribute to a better theoretical understanding of batch normalization.


Persistent Identifierhttp://hdl.handle.net/10722/337341

 

DC FieldValueLanguage
dc.contributor.authorCao, Yuan-
dc.contributor.authorZou, Difan-
dc.contributor.authorLi, Yuanzhi-
dc.contributor.authorGu, Quanquan-
dc.date.accessioned2024-03-11T10:20:07Z-
dc.date.available2024-03-11T10:20:07Z-
dc.date.issued2023-07-15-
dc.identifier.urihttp://hdl.handle.net/10722/337341-
dc.description.abstract<p>We study the implicit bias of batch normalization trained by gradient descent. We show that when learning a linear model with batch normalization for binary classification, gradient descent converges to a uniform margin classifier on the training data with an exp(−Ω(log2t)) convergence rate. This distinguishes linear models with batch normalization from those without batch normalization in terms of both the type of implicit bias and the convergence rate. We further extend our result to a class of two-layer, single-filter linear convolutional neural networks, and show that batch normalization has an implicit bias towards a patch-wise uniform margin. Based on two examples, we demonstrate that patch-wise uniform margin classifiers can outperform the maximum margin classifiers in certain learning problems. Our results contribute to a better theoretical understanding of batch normalization.<br></p>-
dc.languageeng-
dc.relation.ispartof36th Annual Conference on Learning Theory (COLT 2023) (12/07/2023-15/07/2023, Bangalore, India)-
dc.titleThe Implicit Bias of Batch Normalization in Linear Models and Two-layer Linear Convolutional Neural Networks -
dc.typeConference_Paper-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats