Dimitri Meunier
@dimitrimeunier.bsky.social
94 followers 120 following 15 posts
PhD, Gatsby, UCL
Posts Media Videos Starter Packs
Reposted by Dimitri Meunier
emtiyaz.mastodon.social.ap.brid.gy
AISTATS 2026 will be in Morocco!
dimitrimeunier.bsky.social
I have been looking at the draft for a while, I am surprised you had a hard time publishing it, it is a super cool work! Will it be included in the TorchDR package ?
Reposted by Dimitri Meunier
rflamary.bsky.social
Distributional Reduction paper with H. Van Assel, @ncourty.bsky.social, T. Vayer , C. Vincent-Cuaz, and @pfrossard.bsky.social is accepted at TMLR. We show that both dimensionality reduction and clustering can be seen as minimizing an optimal transport loss 🧵1/5. openreview.net/forum?id=cll...
Reposted by Dimitri Meunier
arxiv-stat-ml.bsky.social
Dimitri Meunier, Antoine Moulin, Jakub Wornbard, Vladimir R. Kostic, Arthur Gretton
Demystifying Spectral Feature Learning for Instrumental Variable Regression
https://arxiv.org/abs/2506.10899
dimitrimeunier.bsky.social
Very much looking forward to this ! 🙌 Stellar line-up
lenaicchizat.bsky.social
Announcing : The 2nd International Summer School on Mathematical Aspects of Data Science
mathsdata2025.github.io
EPFL, Sept 1–5, 2025

Speakers:
Bach @bachfrancis.bsky.social
Bandeira
Mallat
Montanari
Peyré @gabrielpeyre.bsky.social

For PhD students & early-career researchers
Apply before May 15!
Mathematical Aspects of Data Science
Graduate Summer School - EPFL - Sept. 1-5, 2025
mathsdata2025.github.io
Reposted by Dimitri Meunier
antoine-mln.bsky.social
new preprint with the amazing @lviano.bsky.social and @neu-rips.bsky.social on offline imitation learning! learned a lot :)

when the expert is hard to represent but the environment is simple, estimating a Q-value rather than the expert directly may be beneficial. lots of open questions left though!
dimitrimeunier.bsky.social
TL;DR:

✅ Theoretical guarantees for nonlinear meta-learning
✅ Explains when and how aggregation helps
✅ Connects RKHS regression, subspace estimation & meta-learning

Co-led with Zhu Li 🙌, with invaluable support from @arthurgretton.bsky.social, Samory Kpotufe.
dimitrimeunier.bsky.social
Even with nonlinear representation you can estimate the shared structure at a rate improving in both N (tasks) and n (samples per task). This leads to parametric rates on the target task!⚡

Bonus: for linear kernels, our results recover known linear meta-learning rates.
dimitrimeunier.bsky.social
Short answer: Yes ✅

Key idea💡: Instead of learning each task well, under-regularise per-task estimators to better estimate the shared subspace in the RKHS.

Even though each task is noisy, their span reveals the structure we care about.

Bias-variance tradeoff in action.
dimitrimeunier.bsky.social
Our paper analyses a meta-learning setting where tasks share a finite dimensional subspace of a Reproducing Kernel Hilbert Space.

Can we still estimate this shared representation efficiently — and learn new tasks fast?
dimitrimeunier.bsky.social
Most prior theory assumes linear structure: All tasks share a linear representation, and task-specific parts are also linear.

Then: we can show improved learning rates as the number of tasks increases.

But reality is nonlinear. What then?
dimitrimeunier.bsky.social
Meta-learning = using many related tasks to help learn new ones faster.

In practice (e.g. with neural nets), this usually means learning a shared representation across tasks — so we can train quickly on unseen ones.

But: what’s the theory behind this? 🤔
Reposted by Dimitri Meunier
arxiv-stat-ml.bsky.social
Dimitri Meunier, Zikai Shen, Mattes Mollenhauer, Arthur Gretton, Zhu Li
Optimal Rates for Vector-Valued Spectral Regularization Learning Algorithms
https://arxiv.org/abs/2405.14778
Reposted by Dimitri Meunier
cslg-bot.bsky.social
Mattes Mollenhauer, Nicole M\"ucke, Dimitri Meunier, Arthur Gretton: Regularized least squares learning with heavy-tailed noise is minimax optimal https://arxiv.org/abs/2505.14214 https://arxiv.org/pdf/2505.14214 https://arxiv.org/html/2505.14214
Reposted by Dimitri Meunier
gabrielpeyre.bsky.social
I have updated my slides on the maths of AI by an optimal pairing between AI and maths researchers ... speakerdeck.com/gpeyre/the-m...
Reposted by Dimitri Meunier
arxiv-stat-ml.bsky.social
Gabriel Peyr\'e
Optimal Transport for Machine Learners
https://arxiv.org/abs/2505.06589
Reposted by Dimitri Meunier
fxbriol.bsky.social
New ICML 2025 paper: Nested expectations with kernel quadrature.

We propose an algorithm to estimate nested expectations which provides orders of magnitude improvements in low-to-mid dimensional smooth nested expectations using kernel ridge regression/kernel quadrature.

arxiv.org/abs/2502.18284
dimitrimeunier.bsky.social
Great talk by Aapo Hyvärinen on non linear ICA at AISTATS 25’!
Reposted by Dimitri Meunier
Density Ratio-based Proxy Causal Learning Without Density Ratios 🤔

at #AISTATS2025

An alternative bridge function for proxy causal learning with hidden confounders.
arxiv.org/abs/2503.08371
Bozkurt, Deaner, @dimitrimeunier.bsky.social, Xu