I am broadly interested in theoretical machine learning, nonparametric statistics and probability theory. The following are some of the topics I am working on right now:
Suppose \(X\) and \(Y\) are \(\mathbb R^d-\)valued random vectors on a probability space \((\Omega, \mathscr F, \mathbb P)\) with distributions \(\mu := \mathbb P \circ X^{-1}\) and \(\nu := \mathbb P \circ Y^{-1}\) respectively. It is often of interest to find, if it exists, a measurable map \(T \colon \mathbb R^d \to \mathbb R^d\) such that the pushforward \(T_{\#} \mu := \mu \circ T^{-1}\) of \(\mu\) under \(T\) equals \(\nu.\) It is possible that no such map exists, for example, if \(\mu = \delta_{x_0}\) for some \(x_0 \in \mathbb R^d\) and \(\nu\) does not equal a Dirac measure, because for any measurable \(T\colon \mathbb R^d \to \mathbb R^d\) the pushforward \(T_{\#}\mu\) is \(\delta_{T(x_0)}.\)
On the other hand, if multiple such maps exist, then which one should we prefer? In the Monge's formulation of optimal transport problem, we have a measurable cost function \(c \colon \mathbb R^d \times \mathbb R^d \to [0, \infty]\), and we want to find an optimal map
In statistical optimal transport, the explicit form of the measures \(\mu\) or \(\nu\) is unknown, and instead only independent samples from them are observed. To that end, let \(X_1, X_2, \ldots\) be independent copies of \(X,\) and let \(Y_1, Y_2, \ldots\) be independent copies of \(Y.\) Let \(\mu_n := \frac{1}{n} \sum_{i=1}^n \delta_{X_i}\) denote the empirical measure based on the sample \((X_1, \ldots, X_n)\) of the first \(n\) observations, and likewise for \(\nu_n.\)
Brenier, Y.: Decomposition polaire et rearrangement monotone des champs de vecteurs. C. R. Acad. Sci. Paris Ser. I Math., 305:805–808, 1987.
Knott, M. and Smith, C. S.: On the optimal mapping of distributions. Journal of Optimization Theory and Applications, 43(1):39–49, 1984.
McCann, R.J.: Existence and uniqueness of monotone measure-preserving maps. Duke Mathematical Journal, 80(2):309 – 323, 1995.
Santambrogio, F.: Optimal Transport for Applied Mathematicians. Progress in Nonlinear Differential Equations and Their Applications, vol. 87. Birkhäuser/Springer, Cham (2015). Calculus of variations, PDEs, and modeling.
Villani, C.: Topics in Optimal Transportation. Graduate Studies in Mathematics, vol. 58. American Mathematical Society, Providence (2003).
NeurIPS 2022, AISTATS 2022, ICLR 2022, NeurIPS 2021, AAAI 2021, ICML 2020
Fall 2021, Spring 2021 — Masters level course on introduction to machine learning