CSML IIT Lab
banner
pontilgroup.bsky.social
CSML IIT Lab
@pontilgroup.bsky.social
560 followers 13 following 28 posts
Computational Statistics and Machine Learning (CSML) Lab | PI: Massimiliano Pontil | Webpage: csml.iit.it | Active research lines: Learning theory, ML for dynamical systems, ML for science, and optimization.
Posts Media Videos Starter Packs
Excited to share our group’s latest work at #AISTATS2025! 🎓
Tackling concentration in dependent data settings with empirical Bernstein bounds for Hilbert space-valued processes.
📍Catch the poster tomorrow!

🔁 See the original tweet for details!
🚨 Poster at #AISTATS2025 tomorrow!
📍Poster Session 1 #125

We present a new empirical Bernstein inequality for Hilbert space-valued random processes—relevant for dependent, even non-stationary data.

w/ Andreas Maurer, @vladimir-slk.bsky.social & M. Pontil

📄 Paper: openreview.net/forum?id=a0E...
DeltaProduct is here! Achieve better state tracing through highly parallel execution. Explore more!🚀
1/9 There is a fundamental tradeoff between parallelizability and expressivity of Large Language Models. We propose a new linear RNN architecture, DeltaProduct, that can effectively navigate this tradeoff. Here's how!
14/ Looking ahead, we’re excited to tackle new challenges:
• Learning from partial observations
• Modeling non-time-homogeneous dynamics
• Expanding applications in neuroscience, genetics, and climate modeling

Stay tuned for groundbreaking updates from our team! 🌍
🙏 Collaborations with the Dynamic Legged Systems group led by Claudio Semini and the Atomistic Simulations group led by Michele Parrinello enriched our research, resulting in impactful works like [P9, P10] and [P7, P11].
12/ This journey wouldn’t have been possible without the inspiring collaborations that shaped our work.

🌟 Special thanks to Karim Lounici from École Polytechnique, whose insights were a major driving force behind many projects.
11/ One of our most exciting results:
[P8] NeurIPS 2024 proposed Neural Conditional Probability (NCP) to efficiently learn conditional distributions. It simplifies uncertainty quantification and guarantees accuracy for nonlinear, high-dimensional data.
10/ [P7] NeurIPS 2024 developed methods to discover slow dynamical modes in systems like molecular simulations. This is transformative for studying rare events and costly data acquisition scenarios in atomistic systems.
9/ Addressing continuous dynamics:
[P6] NeurIPS 2024 introduced a physics-informed framework for learning Infinitesimal Generators (IG) of stochastic systems, ensuring robust spectral estimation.
8/ 🌟 Representation learning takes center stage in:
[P5] ICLR 2024
We combined neural networks with operator theory via Deep Projection Networks (DPNets). This approach enhances robustness, scalability, and interpretability for dynamical systems.
7/ 📈 Scaling up:
[P4] NeurIPS 2023 introduced a Nyström sketching-based method to reduce computational costs from cubic to almost linear without sacrificing accuracy. Validated on massive datasets like molecular dynamics, see figure.
6/ [P3] ICML 2024 addressed a critical issue in TO-based modeling: reliable long-term predictions.
Our Deflate-Learn-Inflate (DLI) paradigm ensures uniform error bounds, even for infinite time horizons. This method stabilized predictions in real-world tasks; see the figure.
5/ [P2] NeurIPS 2023 advanced TOs with theoretical guarantees for spectral decomposition—previously lacking finite sample guarantees. We developed sharp learning rates, enabling accurate, reliable models for long-term system behavior.
4/ 🔑 The journey began with:
[P1] NeurIPS 2022
We introduced the first ML formulation for learning TO, which led to the development of the open-source Kooplearn library. This step laid the groundwork for exploring the theoretical limits of operator learning from finite data.
3/TOs describe system evolution over finite time intervals, while IGs capture instantaneous rates of change. Their spectral decomposition is key for identifying dominant modes and understanding long-term behavior in complex or stochastic systems.
2/ 🌐 Our work revolves around Markov/Transfer Operators (TO) and their Infinitesimal Generators (IG)—tools that allow us to model complex dynamical systems by understanding their evolution in higher-dimensional spaces. Here’s why this matters.
1/ 🚀 Over the past two years, our team, CSML, at IIT, has made significant strides in the data-driven modeling of dynamical systems. Curious about how we use advanced operator-based techniques to tackle real-world challenges? Let’s dive in! 🧵👇
An inspiring dive into understanding dynamical processes through 'The Operator Way.' A fascinating approach made accessible for everyone—check it out! 👇👀
For the past four years, I’ve been working on a topic that’s both fascinating and challenging to explain. In this post, I’ve tried to present The Operator Way — a paradigm for understanding dynamical processes — in plain, approachable terms.

pietronvll.github.io/the-operator...
Reposted by CSML IIT Lab
Excited to present
"Unlocking State-Tracking in Linear RNNs Through Negative Eigenvalues"
at the M3L workshop at #NeurIPS
https://buff.ly/3BlcD4y

If interested, you can attend the presentation the 14th at 15:00, pass at the afternoon poster session, or DM me to discuss :)
Unlocking State-Tracking in Linear RNNs Through Negative Eigenvalues
Linear Recurrent Neural Networks (LRNNs) such as Mamba, RWKV, GLA, mLSTM, and DeltaNet have emerged as efficient alternatives to Transformers in large language modeling, offering linear scaling with…
buff.ly
Reposted by CSML IIT Lab
In his book “The Nature of Statistical Learning” V. Vapnik wrote:
“When solving a given problem, try to avoid a more general problem as an intermediate step”
Join us at our posters and talks to connect, share ideas, and explore collaborations. 🚀✨
🔬 Fine-tuning Foundation Models for Molecular Dynamics: A Data-Efficient Approach with Random Features
✍️ @pienovelli.bsky.social, L. Bonati, P. Buigues, G. Meanti, L. Rosasco, M. Pontil | 📅ML4PS Workshop, Dec 15.
🔗 Unlocking State-Tracking in Linear RNNs Through Negative Eigenvalues
✍️ R. Grazzi, J. Siems, J. Franke, A. Zela, F. Hutter, M. Pontil
📃https://arxiv.org/abs/2411.12537 | 📅 Oral @ M3L workshop, Dec 14, 15:00 - 15:15.
🌊 Learning the Infinitesimal Generator of Stochastic Diffusion Processes
✍️V. Kostic, H. Halconruy, @tdevergne.bsky.social, K. Lounici, M. Pontil
📃https://arxiv.org/abs/2405.12940 | 📅 Poster #5410 Dec 13, 16:30 - 19:30.