Show simple item record

dc.contributor.advisorJegelka, Stefanie
dc.contributor.authorLim, Derek
dc.date.accessioned2026-04-21T20:43:16Z
dc.date.available2026-04-21T20:43:16Z
dc.date.issued2025-09
dc.date.submitted2025-09-15T14:41:17.184Z
dc.identifier.urihttps://hdl.handle.net/1721.1/165587
dc.description.abstractModern neural networks are large, complex objects, which can be difficult to study and work with. In this thesis, I analyze and improve neural networks from the perspective of symmetries, with particular focus on function symmetries and parameter symmetries. Function symmetries are transformations of the input that lead to predictable changes in the output, which can be enforced in neural network architectures to improve performance on data with symmetry structures. Parameter symmetries are transformations of parameters that leave the underlying neural network function unchanged, and they have impacts on various empirical phenomena in neural networks. In Part I of this thesis, I focus on function symmetries, and develop new methods and analysis techniques for equivariant neural networks that have function symmetries baked into their architectures. I apply these techniques primarily on eigenvector-valued data, resulting in the first provably expressive neural network architectures that respect the symmetries of eigenvector data. In Part II, I focus on parameter-symmetries, and analyze their impact in various empirical phenomena of neural networks, as well as their impact in the open-weight ecosystem of models with publicly-shared parameters. In Part III, I consider both function and parameter symmetries to construct metanetworks: models that take in the parameters of other neural networks as input. Since the input to metanetworks are parameters, I develop metanetworks that are invariant or equivariant to the parameter symmetries of the input networks. All in all, my work shows that accounting for function and parameter symmetries is both theoretically and empirically beneficial across diverse types of data, learning tasks, neural network architectures, and other parts of the deep learning pipeline.
dc.publisherMassachusetts Institute of Technology
dc.rightsIn Copyright - Educational Use Permitted
dc.rightsCopyright retained by author(s)
dc.rights.urihttps://rightsstatements.org/page/InC-EDU/1.0/
dc.titleSymmetries in Neural Network Functions and Parameters
dc.typeThesis
dc.description.degreePh.D.
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Science
mit.thesis.degreeDoctoral
thesis.degree.nameDoctor of Philosophy


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record