TY - JOUR
T1 - OPTIMAL TRANSPORT MAP ESTIMATION IN GENERAL FUNCTION SPACES
AU - Divol, Vincent
AU - Niles-Weed, Jonathan
AU - Pooladian, Aram Alexandre
N1 - Publisher Copyright:
© Institute of Mathematical Statistics, 2025.
PY - 2025/6/1
Y1 - 2025/6/1
N2 - We study the problem of estimating a function T, given independent samples from a distribution P and from the pushforward distribution T♯ P. This setting is motivated by applications in the sciences, where T represents the evolution of a physical system over time, and in machine learning, where, for example, T may represent a transformation learned by a deep neural network trained for a generative modeling task. To ensure identifiability, we assume that T = ∇φ0 is the gradient of a convex function in which case T is known as an optimal transport map. Prior work has studied the estimation of T under the assumption that it lies in a Hölder class, but general theory is lacking. We present a unified methodology for obtaining rates of estimation of optimal transport maps in general function spaces. Our assumptions are significantly weaker than those appearing in the literature: we require only that the source measure P satisfy a Poincaré inequality and that the optimal map be the gradient of a smooth convex function that lies in a space whose metric entropy can be controlled. As a special case, we recover known estimation rates for Hölder transport maps but also obtain nearly sharp results in many settings not covered by prior work. For example, we provide the first statistical rates of estimation when P is the normal distribution and the transport map is given by an infinite-width shallow neural network.
AB - We study the problem of estimating a function T, given independent samples from a distribution P and from the pushforward distribution T♯ P. This setting is motivated by applications in the sciences, where T represents the evolution of a physical system over time, and in machine learning, where, for example, T may represent a transformation learned by a deep neural network trained for a generative modeling task. To ensure identifiability, we assume that T = ∇φ0 is the gradient of a convex function in which case T is known as an optimal transport map. Prior work has studied the estimation of T under the assumption that it lies in a Hölder class, but general theory is lacking. We present a unified methodology for obtaining rates of estimation of optimal transport maps in general function spaces. Our assumptions are significantly weaker than those appearing in the literature: we require only that the source measure P satisfy a Poincaré inequality and that the optimal map be the gradient of a smooth convex function that lies in a space whose metric entropy can be controlled. As a special case, we recover known estimation rates for Hölder transport maps but also obtain nearly sharp results in many settings not covered by prior work. For example, we provide the first statistical rates of estimation when P is the normal distribution and the transport map is given by an infinite-width shallow neural network.
KW - Optimal transport
KW - nonparametric estimation
UR - https://www.scopus.com/pages/publications/105010285119
U2 - 10.1214/24-AOS2482
DO - 10.1214/24-AOS2482
M3 - Article
AN - SCOPUS:105010285119
SN - 0090-5364
VL - 53
SP - 963
EP - 988
JO - Annals of Statistics
JF - Annals of Statistics
IS - 3
ER -