Show simple item record

dc.contributor.advisorMarzouk, Youssef
dc.contributor.authorRen, Zhi (Robert)
dc.date.accessioned2025-07-07T17:36:59Z
dc.date.available2025-07-07T17:36:59Z
dc.date.issued2025-05
dc.date.submitted2025-06-05T14:18:35.570Z
dc.identifier.urihttps://hdl.handle.net/1721.1/159886
dc.description.abstractSampling from an arbitrary probability distribution is a central problem in computational statistics and machine learning. Transportation of measure offers a useful approach to this problem: the idea is to construct a measurable map that pushes forward a relatively simple source distribution to the target probability distribution. One can then simulate from the target distribution by drawing samples from the source distribution and evaluating the transport map. This construction is applicable to both generative modeling and variational inference; when the map is invertible, one can also estimate the density of the target measure by evaluating the density of the pushforward of the source distribution under the inverse transport map. Over the past decade, various parameterizations of such transports have been proposed. Generally speaking, they fall into two categories: the static approach, where the displacement from x to T(x) is represented directly and the dynamic approach that employs evolution of measures by some differential equation over some fictitious time. While many of these models have achieved enormous success in practical applications, their theoretical underpinnings remain largely unexplored. In this thesis, we provide a theoretical foundation for flow-based methods for sampling and generative modeling, and a unified view of both continuous and discrete-time approaches. In the first part of the thesis, we address the approximation theory of flow based methods. In particular, we show how the regularity of the underlying ODE velocity field relates to the regularity of densities and prove related neural network approximation bounds. In addition, we show how introduction of a time-reparameterized schedule can dramatically improve the regularity of the velocity, helping resolve potential singularities. In the second part of the thesis, we focus on the interplay between flow-based models and nonparametric statistics. In particular, we consider pullback density estimators under these flow based models from likelihood-based objectives. The estimators we consider arise from both discrete and continuous-time parameterizations of the transport, and the underlying function classes we consider include Hölder balls and neural networks. In all these cases, we show they achieve near minimax optimal rate for learning s-smooth densities.
dc.publisherMassachusetts Institute of Technology
dc.rightsIn Copyright - Educational Use Permitted
dc.rightsCopyright retained by author(s)
dc.rights.urihttps://rightsstatements.org/page/InC-EDU/1.0/
dc.titleTheoretical Foundations of Flow-based Methods for Sampling and Generative Modeling
dc.typeThesis
dc.description.degreePh.D.
dc.contributor.departmentMassachusetts Institute of Technology. Department of Mathematics
dc.contributor.departmentMassachusetts Institute of Technology. Center for Computational Science and Engineering
mit.thesis.degreeDoctoral
thesis.degree.nameDoctor of Philosophy


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record