Show simple item record

dc.contributor.advisorTorralba, Antonio
dc.contributor.authorBau, David
dc.date.accessioned2022-02-07T15:20:58Z
dc.date.available2022-02-07T15:20:58Z
dc.date.issued2021-09
dc.date.submitted2021-09-21T19:30:13.523Z
dc.identifier.urihttps://hdl.handle.net/1721.1/140048
dc.description.abstractWe investigate the role of neurons within the internal computations of deep neural networks for computer vision. We introduce network dissection, a method for quantifying the alignment between human-interpretable visual concepts and individual neurons in a deep network. We apply network dissection to examine and compare the internal computations of several networks trained to classify and represent images, and we ask how well human-understandable concepts align with neurons at different layers, in different architectures, with various training objectives; we also compare neurons to random linear combinations of neurons, and examine emergence of concepts as training proceeds. Then, we adapt network dissection to analyze generative adversarial networks. In GAN dissection, human-understandable neurons are identified by applying a semantic segmentation model to generated output. We find that small sets of neurons control the presence of specific objects within synthesized scenes. We also find that activating neurons reveals modeled rules and interactions between objects and their context. We then ask how to dissect and understand the omissions of a generative network. Omissions of human-understandable objects can be quantified by comparing semantic segmentation statistics between the training distribution and the generated distribution. Then we develop a method that can invert and reconstruct generated images in a progressive GAN, and show that this reconstruction can visualize specific cases in which the GAN omits identified object classes. Finally, we ask how rules within a generative model are represented. We hypothesize that the layers of a generative model serve as a memory that stores associations from representations of concepts at the input of a layer to patterns of concepts at the output to the layer, and we develop a method for rewriting the weights of a model by directly rewriting one memorized association. We show that our method can be used to rewrite several individual associative memories in a Progressive GAN or StyleGAN, altering learned rules that govern the appearance of specific object parts in the model.
dc.publisherMassachusetts Institute of Technology
dc.rightsIn Copyright - Educational Use Permitted
dc.rightsCopyright MIT
dc.rights.urihttp://rightsstatements.org/page/InC-EDU/1.0/
dc.titleDissection of Deep Neural Networks
dc.typeThesis
dc.description.degreePh.D.
dc.contributor.departmentMassachusetts Institute of Technology. Department of Electrical Engineering and Computer Science
dc.identifier.orcidhttps://orcid.org/0000-0003-1744-6765
mit.thesis.degreeDoctoral
thesis.degree.nameDoctor of Philosophy


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record