Show simple item record

dc.contributor.authorPinto, Andrea
dc.contributor.authorRangamani, Akshay
dc.contributor.authorPoggio, Tomaso
dc.date.accessioned2024-10-11T13:51:12Z
dc.date.available2024-10-11T13:51:12Z
dc.date.issued2024-10-11
dc.identifier.urihttps://hdl.handle.net/1721.1/157263
dc.description.abstractWhile previous optimization results have suggested that deep neural networks tend to favour low-rank weight matrices, the implications of this inductive bias on generalization bounds remain under-explored. In this paper, we apply a chain rule for Gaussian complexity (Maurer, 2016a) to analyze how low-rank layers in deep networks can prevent the accumulation of rank and dimensionality factors that typically multiply across layers. This approach yields generalization bounds for rank and spectral norm constrained networks. We compare our results to prior generalization bounds for deep networks, highlighting how deep networks with low-rank layers can achieve better generalization than those with full-rank layers. Additionally, we discuss how this framework provides new perspectives on the generalization capabilities of deep nets exhibiting neural collapse.en_US
dc.description.sponsorshipThis work was supported by the Center for Brains, Minds and Machines (CBMM), funded by NSF STC award CCF-1231216.en_US
dc.publisherCenter for Brains, Minds and Machines (CBMM)en_US
dc.relation.ispartofseriesCBMM Memo;151
dc.subjectGaussian Complexity, Generalization Bounds, Low Rank Layers, Neural Collapseen_US
dc.titleOn Generalization Bounds for Neural Networks with Low Rank Layersen_US
dc.typeArticleen_US
dc.typeTechnical Reporten_US
dc.typeWorking Paperen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record