Arthur Gretton

461 Followers
37 Following
3 Posts
Professor, Gatsby Computational Neuroscience Unit

Optimal Rates for Regularized Conditional Mean Embedding Learning
"what it says on the tin"😁

Oral presentation #NeurIPS22

arXiv: https://arxiv.org/abs/2208.01711
short video: https://youtu.be/Pl8OM2sckwA
Poster #838 Hall J Thursday 01 Dec 4:30
Zhu Li, Dimitri Meunier, Mattes Mollenhauer

Optimal Rates for Regularized Conditional Mean Embedding Learning

We address the consistency of a kernel ridge regression estimate of the conditional mean embedding (CME), which is an embedding of the conditional distribution of $Y$ given $X$ into a target reproducing kernel Hilbert space $\mathcal{H}_Y$. The CME allows us to take conditional expectations of target RKHS functions, and has been employed in nonparametric causal and Bayesian inference. We address the misspecified setting, where the target CME is in the space of Hilbert-Schmidt operators acting from an input interpolation space between $\mathcal{H}_X$ and $L_2$, to $\mathcal{H}_Y$. This space of operators is shown to be isomorphic to a newly defined vector-valued interpolation space. Using this isomorphism, we derive a novel and adaptive statistical learning rate for the empirical CME estimator under the misspecified setting. Our analysis reveals that our rates match the optimal $O(\log n / n)$ rates without assuming $\mathcal{H}_Y$ to be finite dimensional. We further establish a lower bound on the learning rate, which shows that the obtained upper bound is optimal.

arXiv.org