On Generalization Bounds for Neural Networks with Low Rank Layers

TitleOn Generalization Bounds for Neural Networks with Low Rank Layers
Publication TypeCBMM Memos
Year of Publication2024
AuthorsPinto, A, Rangamani, A, Poggio, T
Number151
Date Published10/2024
KeywordsGaussian Complexity, Generalization Bounds, Low Rank Layers, Neural Collapse
Abstract

While previous optimization results have suggested that deep neural networks tend to favour low-rank weight matrices, the implications of this inductive bias on generalization bounds remain under-explored. In this paper, we apply a chain rule for Gaussian complexity (Maurer, 2016a) to analyze how low-rank layers in deep networks can prevent the accumulation of rank and dimensionality factors that typically multiply across layers. This approach yields generalization bounds for rank and spectral norm constrained networks. We compare our results to prior generalization bounds for deep networks, highlighting how deep networks with low-rank layers can achieve better generalization than those with full-rank layers. Additionally, we discuss how this framework provides new perspectives on the generalization capabilities of deep nets exhibiting neural collapse.

DSpace@MIT

https://hdl.handle.net/1721.1/157263

Download:  PDF icon CBMM-Memo-151.pdf
CBMM Memo No:  151

Associated Module: 

CBMM Relationship: 

  • CBMM Funded