Maarten van Smeden
@maartenvsmeden.bsky.social
10K followers 480 following 290 posts
statistician • associate prof • team lead health data science and head methods research program at julius center • director ai methods lab, umc utrecht, netherlands • views and opinions my own
Posts Media Videos Starter Packs
maartenvsmeden.bsky.social
Kind reminder: data driven variable selection (e.g. forward/stepwise/univariable screening) makes things *worse* for most analytical goals
Reposted by Maarten van Smeden
statsepi.bsky.social
Interpretable "AI" is just a distraction from safe and useful "AI"
maartenvsmeden.bsky.social
This is right tho. Let’s therefore call them sensitivity positive predictive value curves bsky.app/profile/laur...
lauretig.bsky.social
9. It's annoying how often the same model is "discovered" in a different field, with a completely different set of jargon
maartenvsmeden.bsky.social
No.
lauretig.bsky.social
5. You should use a precision-recall curve for a binary classifier, not an ROC curve
maartenvsmeden.bsky.social
I wonder who those people are who come here dying to know what GenAI has done with some prompt you put in
maartenvsmeden.bsky.social
If you think AI is cool, wait until you learn about regression analysis
maartenvsmeden.bsky.social
TL;DR: Explainable AI models often don't do a good job explaining. They can be very useful for description. We should be really careful when using Explainable AI in clinical decision making, and even when judging face validity of AI models

Excellently led by @alcarriero.bsky.social
maartenvsmeden.bsky.social
NEW PREPRINT

Explainable AI refers to an extremely popular group of approaches that aim to open "black box" AI models. But what can we see when we open the black AI box? We use Galit Shmueli's framework (to describe, predict or explain) to evaluate

arxiv.org/abs/2508.05753
maartenvsmeden.bsky.social
The healthcare literature is filled with "risk factors". This word combination makes research findings sound important by implying causality, while avoiding direct claims of having identified causal associations that are easily critiqued.
maartenvsmeden.bsky.social
And taking this analogy one step further: it gives genuine phone repair shops a bad name
maartenvsmeden.bsky.social
When forced to make a choice, my choice will be logistic regression model over linear probability model 103% of the time
Reposted by Maarten van Smeden
timpmorris.bsky.social
Post just up: Is multiple imputation making up information?

tldr: no.

Includes a cheeky simulation study to demonstrate the point.
open.substack.com/pub/tpmorris...
Cover picture with blog title & subtitle, and results graph in the background
Reposted by Maarten van Smeden
statsepi.bsky.social
You can have all the omni-omics data in the world and the bestest algorithms, but eventually a predicted probability is produced & it should be evaluated using well-established methods, and correctly implemented in the context of medical decision making.

statsepi.substack.com/i/140315566/...
The leaky pipe of clinical prediction models. by @maartenvsmeden.bsky.social‬ et al
maartenvsmeden.bsky.social
Clients: “I want to find real, meaningful clusters”
Me: “I want world peace, which is more likely to happen than what you want”
maartenvsmeden.bsky.social
Depending which methods guru you ask every analytical task is “essentially” a missing data problem, a causal inference problem, a Bayesian problem, a regression problem or a machine learning problem
Reposted by Maarten van Smeden
maartenvsmeden.bsky.social
In medicine they are called "risk factors" and, of course, you want all "important" risk factors in your model all the time

Unless a risk factor is not statistically significant then you can drop that factor without issues
Reposted by Maarten van Smeden
richarddriley.bsky.social
* New preprint led by Joao Matos & @gscollins.bsky.social

"Critical Appraisal of Fairness Metrics in Clinical Predictive AI"

- Important, rapidly growing area
- But confusion exists
- 62 fairness metrics identified so far
- Better standards & metrics needed for healthcare
arxiv.org/abs/2506.17035
maartenvsmeden.bsky.social
Also, the fact that a model with the best AUC doesn't always mean the model makes the best predictions is lost in such cases too
maartenvsmeden.bsky.social
Surprisingly common thing: comparisons of prediction models developed using, say, Logistic Regression, Random Forest and XGBoost with conclusion XGBoost is "good" because it yields slightly higher AUC than LR or RF using the same data

Fact that "better" doesn't always mean "good" seems lost
Reposted by Maarten van Smeden
georgheinze.bsky.social
Published: the paper 'On the uses and abuses of Regression Models: a Call for Reform of Statistical Practice and Teaching' by John Carlin and Margarita Moreno-Betancur in the latest issue of Statistics in Medicine onlinelibrary.wiley.com/doi/10.1002/... (1/8)
onlinelibrary.wiley.com
maartenvsmeden.bsky.social
What is common knowledge in your field, but shocks outsiders?

Validated does not mean it works as intended. It means someone has evaluated it (and may have concluded it doesn’t work at all)
editoratlarge.bsky.social
What is common knowledge in your field, but shocks outsiders?

We're not clear on what peer review is, at all.
jensfoell.de
What is common knowledge in your field, but shocks outsiders?

We’re not clear on what intelligence is, at all