Lennart Purucker
@lennartpurucker.bsky.social
130 followers 150 following 16 posts
PhD student sup. by Frank Hutter; researching automated machine learning and foundation models for (small) tabular data! Website: https://ml.informatik.uni-freiburg.de/profile/purucker/
Posts Media Videos Starter Packs
Pinned
lennartpurucker.bsky.social
🚨What is SOTA on tabular data, really? We are excited to announce 𝗧𝗮𝗯𝗔𝗿𝗲𝗻𝗮, a living benchmark for machine learning on IID tabular data with:

📊 an online leaderboard (submit!)
📑 carefully curated datasets
📈 strong tree-based, deep learning, and foundation models

🧵
Reposted by Lennart Purucker
madelonhulsebos.bsky.social
🚀 Excited to announce the AI for Tabular Data workshop at EurIPS 2025 in Copenhagen!

CfP: sites.google.com/view/eurips2... (papers due 20 Oct)

Join us @euripsconf.bsky.social to discuss neural tabular models and systems for predictive ML, tabular reasoning and retrieval, table synthesis and more ✨
Announcement for the EurIPS'25 Workshop on AI for Tabular Data. More information about the workshop can be found at: https://sites.google.com/view/eurips25-ai-td/home.
Reposted by Lennart Purucker
matthiasfeurer.bsky.social
Very proud our paper "OpenML: Insights from 10 years and more than a thousand papers" that describes the current state of OpenML and also analyzes the impact OpenML had over the last years (yes, we manually looked at every paper citing OpenML). #openml #automl #openscience 1/4
lennartpurucker.bsky.social
TabArena is a living benchmark. With the community, we will continually update it!

Authors: @nickerickson.bsky.social Lennart Purucker @atschalz.bsky.social @dholzmueller.bsky.social Prateek Desai David Salinas Frank Hutter
LB: tabarena.ai
Paper: arxiv.org/abs/2506.16791
Code: tabarena.ai/code
12/
lennartpurucker.bsky.social
The TabArena team consists of experienced researchers and open-source developers. At the same time, we are also authors of some of the methods benchmarked in our work. We challenge you to find any mistakes or biases in our work to further improve TabArena!

11/
lennartpurucker.bsky.social
We are continuing to improve the TabArena and its usability. You can already use our implementations of all models we benchmarked with scikit-learn-like interfaces:

10/
lennartpurucker.bsky.social
Many benchmarks evaluate methods using holdout validation. We show that this incorrectly represents the relative comparison of methods and results in much worse peak performance! Non-ensemble methods like RealMLP and ModernNCA gain more from 8-fold CV.

9/
lennartpurucker.bsky.social
The worth of models lies not only in their individual performance but also in their contribution to a multi-model ensemble. We build an ensemble of strong and diverse model configurations and show that it significantly outperforms the current SOTA on tabular data, AutoGluon.

8/
lennartpurucker.bsky.social
In terms of training and inference time, tree-based methods still shine compared to modern neural networks.

7/
lennartpurucker.bsky.social
We evaluate three foundation models. TabDPT runs on every dataset and is mid-field with good regression results. TabPFNv2 and TabICL achieve very good results on subsets of the benchmark within their corresponding dataset constraints (left: TabPFNv2, right: TabICL).

6/
lennartpurucker.bsky.social
On the full benchmark, the recent deep learning models RealMLP and TabM take the top spots with weighted ensembling, slightly outperforming boosted trees on average, although boosted trees are faster. Without ensembling, CatBoost wins.

5/
lennartpurucker.bsky.social
Where possible, we coordinate with authors to obtain good hyperparameter search spaces. For tree-based baselines, we took implementations from AutoGluon and made them better by carefully optimizing their search spaces, so these might be the best baselines out there.

4/
lennartpurucker.bsky.social
We curated datasets by 𝗺𝗮𝗻𝘂𝗮𝗹𝗹𝘆 checking 1053 datasets from prior benchmarks. Only 51 were realistic tabular IID predictive tasks with 500-250K samples, which we share via OpenML. Together with the community, we aim to extend TabArena's datasets in the future!

3/
lennartpurucker.bsky.social
TabArena implements best practices for SOTA performance: 8-fold inner cross-validation with bagging, outer cross-validation for evaluation, early stopping where possible, extensive tuning, and weighted ensembles of hyperparameter configurations to obtain peak performance.

2/
lennartpurucker.bsky.social
🚨What is SOTA on tabular data, really? We are excited to announce 𝗧𝗮𝗯𝗔𝗿𝗲𝗻𝗮, a living benchmark for machine learning on IID tabular data with:

📊 an online leaderboard (submit!)
📑 carefully curated datasets
📈 strong tree-based, deep learning, and foundation models

🧵
Reposted by Lennart Purucker
nickerickson.bsky.social
We are excited to announce #FMSD: "1st Workshop on Foundation Models for Structured Data" has been accepted to #ICML 2025!
Call for Papers: icml-structured-fm-workshop.github.io
lennartpurucker.bsky.social
The tabular foundation model TabPFN v2 is finally public 🎉🥳
This is excellent news for (small) tabular ML! Checkout our Nature article (nature.com/articles/s41...) and code (github.com/PriorLabs/Ta...)
lennartpurucker.bsky.social
Excited to be at #NeurIPS2024 tomorrow! 🎉 Let’s connect if you are interested in tabular data and:

🤖 AutoML (e.g., AutoGluon)
📊 Data Science (e.g., LLMs for Feature Engineering)
🏛️ Foundation Models (e.g., TabPFN)

Looking forward to insightful discussions—feel free to reach out!
lennartpurucker.bsky.social
Mhm, I think arxiv.org/pdf/2403.01554 might be an interesting starting point. But I also do not have a good overview right now.
arxiv.org
lennartpurucker.bsky.social
AutoGluon 1.2 is now public! Some highlights for tabular data:
1. 70% win-rate vs version 1.1.
2. TabPFNMix foundation model, parallel fit
3. AutoGluon-Assistant: Zero-code ML with LLMs.

Also, very cool improvements for TimeSeries related to Chronos!

#AutoML #TabularData
Release v1.2.0 · autogluon/autogluon
Version 1.2.0 We're happy to announce the AutoGluon 1.2.0 release. AutoGluon 1.2 contains massive improvements to both Tabular and TimeSeries modules, each achieving a 70% win-rate vs AutoGluon 1.1...
github.com
Reposted by Lennart Purucker
caroladoerr.bsky.social
Okay, okay, @theeimer.bsky.social , here you go: the beginning of an #AutoML starter pack
-->happy to add AutoML folks and friends, just reply/PN to suggest yourself or others
go.bsky.app/5PnMDUK