Show simple item record

dc.contributor.advisorLeslie Kaelbling
dc.contributor.authorKawaguchi, Kenjien_US
dc.contributor.authorKaelbling, Leslie Packen_US
dc.contributor.authorBengio, Yoshuaen_US
dc.contributor.otherLearning and Intelligent Systemsen
dc.date.accessioned2018-05-09T19:55:51Z
dc.date.available2018-05-09T19:55:51Z
dc.date.issued2018-05-01
dc.identifier.urihttp://hdl.handle.net/1721.1/115274
dc.description.abstractWith a direct analysis of neural networks, this paper presents a mathematically tight generalization theory to partially address an open problem regarding the generalization of deep learning. Unlike previous bound-based theory, our main theory is quantitatively as tight as possible for every dataset individually, while producing qualitative insights competitively. Our results give insight into why and how deep learning can generalize well, despite its large capacity, complexity, possible algorithmic instability, nonrobustness, and sharp minima, answering to an open question in the literature. We also discuss limitations of our results and propose additional open problems.en_US
dc.format.extent31 pagesen_US
dc.relation.ispartofseriesMIT-CSAIL-TR-2018-014
dc.rightsCreative Commons Attribution 4.0 International*
dc.rights.urihttp://creativecommons.org/licenses/by/4.0/en_US
dc.subjectneural networken_US
dc.subjectlearning theoryen_US
dc.titleGeneralization in Deep Learningen_US
dc.date.updated2018-05-09T19:55:51Z


Files in this item

Thumbnail
Thumbnail

This item appears in the following Collection(s)

Show simple item record