Bio: Meyer Scetbon is currently a Research Scientist at Microsoft Research. He completed his PhD at Institut Polytechnique de Paris, advised by M. Cuturi. He did, as a visiting student, his MS theses UW and Technion, on kernel-based viewpoints on deep neural networks advised by Z. Harchaoui, and end-to-end signal and image denoising advised by M. Elad, respectively.
Abstract: https://meyerscetbon.github.io/_pages/publications/>
Optimal transport (OT) plays an increasingly important role in machine learning (ML) to compare probability distributions. Yet, it poses, in its original form, several challenges when used for applied problems: (i) computing OT between discrete distributions amounts to solving a large and expensive network flow problem which requires a supercubic complexity in the number of points; (ii) estimating OT using sampled measures is doomed by the curse of dimensionality. These issues can be mitigated using an entropic regularization, solved with the Sinkhorn algorithm, which improves on both statistical and computational aspects. While much faster, entropic OT still requires a quadratic complexity with respect to the number of points and therefore remains prohibitive for large-scale problems. In this talk, I will present new regularization approaches for the OT problem, as well as its quadratic extension, the Gromov-Wasserstein (GW) problem, which impose low-rank structures on the admissible couplings. This results in the development of new algorithms that enjoy a linear complexity both in time and memory with respect to the number of points, enabling their applications in the large-scale setting where millions of points need to be compared. Additionally, I will show that these new regularization schemes have better statistical performances compared to the entropic approach, that they naturally interpolate between the Maximum Mean Discrepancy (MMD) and OT, and that they offer general clustering methods for arbitrary geometry. Website: <