Maarten van Smeden
@maartenvsmeden.bsky.social
10K followers 480 following 290 posts

statistician • associate prof • team lead health data science and head methods research program at julius center • director ai methods lab, umc utrecht, netherlands • views and opinions my own

Mathematics 52%
Public Health 16%
Posts Media Videos Starter Packs

maartenvsmeden.bsky.social
Kind reminder: data driven variable selection (e.g. forward/stepwise/univariable screening) makes things *worse* for most analytical goals

maartenvsmeden.bsky.social
This is right tho. Let’s therefore call them sensitivity positive predictive value curves bsky.app/profile/laur...
lauretig.bsky.social
9. It's annoying how often the same model is "discovered" in a different field, with a completely different set of jargon

maartenvsmeden.bsky.social
No.
lauretig.bsky.social
5. You should use a precision-recall curve for a binary classifier, not an ROC curve

maartenvsmeden.bsky.social
I wonder who those people are who come here dying to know what GenAI has done with some prompt you put in

maartenvsmeden.bsky.social
If you think AI is cool, wait until you learn about regression analysis

maartenvsmeden.bsky.social
TL;DR: Explainable AI models often don't do a good job explaining. They can be very useful for description. We should be really careful when using Explainable AI in clinical decision making, and even when judging face validity of AI models

Excellently led by @alcarriero.bsky.social

maartenvsmeden.bsky.social
NEW PREPRINT

Explainable AI refers to an extremely popular group of approaches that aim to open "black box" AI models. But what can we see when we open the black AI box? We use Galit Shmueli's framework (to describe, predict or explain) to evaluate

arxiv.org/abs/2508.05753

maartenvsmeden.bsky.social
The healthcare literature is filled with "risk factors". This word combination makes research findings sound important by implying causality, while avoiding direct claims of having identified causal associations that are easily critiqued.
lauretig.bsky.social
9. It's annoying how often the same model is "discovered" in a different field, with a completely different set of jargon

Reposted by Maarten van Smeden

lauretig.bsky.social
5. You should use a precision-recall curve for a binary classifier, not an ROC curve

Reposted by Maarten van Smeden

carlbergstrom.com
Wait people are sending MDPI cash money?
Bar graph. 

Figure 4. Estimate of annual APC revenue (in millions USD) by publisher and OA type adjusted for inflation to 2023 USD using CPI Advanced Economies

Reposted by Maarten van Smeden

jeremymberg.bsky.social
Once my selection had been approved by the AAAS board, I spoke with Rush Holt about details. He said that the salary would be the same as the current EiC Marcia McNutt, namely $500, 000/year. I was surprised and, frankly, a bit confused.

10/n
a large pile of gold coins is being poured out of a vault
ALT: a large pile of gold coins is being poured out of a vault
media.tenor.com
maartenvsmeden.bsky.social
Periodic reminder the world of data analysis cannot be meaningfully categorised into "machine learning" and "statistics". Two cultures with substantial overlap in the use of methods (e.g. logistic regression), analytical goals (e.g. causal inference) and history

jamanetwork.com/journals/jam...
eikofried.bsky.social
Wrote Scientific Reports February 8 2024 that a newly published meta-analysis on mindfulness & brain morphology excluded all null-findings and therefore ... by definition found a relationship.

Still no proper response from the journal (other then many "we'll look into it"). It's been a year now.

Reposted by Maarten van Smeden

Reposted by Maarten van Smeden

statsepi.bsky.social
I'm now audience captured. A few more gems:
A bar chart (fig 3) showing the proportion of observations falling into one of two categories was 0.5 for both. Caption: To explore further and make a more balanced dataset for training the models, we have also used SMOTE oversampling technique to resample the dataset and make it a 1:1 ration. Fig 3 shows the training dataset after applying SMOTE oversampling technique.

Reposted by Maarten van Smeden

editoratlarge.bsky.social
What is common knowledge in your field, but shocks outsiders?

We're not clear on what peer review is, at all.
jensfoell.de
What is common knowledge in your field, but shocks outsiders?

We’re not clear on what intelligence is, at all
vortexegg.com
What is common knowledge in your field, but shocks outsiders?

We’re not clear on what *information* is, at all

maartenvsmeden.bsky.social
And taking this analogy one step further: it gives genuine phone repair shops a bad name

maartenvsmeden.bsky.social
When forced to make a choice, my choice will be logistic regression model over linear probability model 103% of the time

Reposted by Maarten van Smeden

timpmorris.bsky.social
Post just up: Is multiple imputation making up information?

tldr: no.

Includes a cheeky simulation study to demonstrate the point.
open.substack.com/pub/tpmorris...
Cover picture with blog title & subtitle, and results graph in the background

Reposted by Maarten van Smeden

statsepi.bsky.social
You can have all the omni-omics data in the world and the bestest algorithms, but eventually a predicted probability is produced & it should be evaluated using well-established methods, and correctly implemented in the context of medical decision making.

statsepi.substack.com/i/140315566/...
The leaky pipe of clinical prediction models. by @maartenvsmeden.bsky.social‬ et al

maartenvsmeden.bsky.social
Clients: “I want to find real, meaningful clusters”
Me: “I want world peace, which is more likely to happen than what you want”

maartenvsmeden.bsky.social
Depending which methods guru you ask every analytical task is “essentially” a missing data problem, a causal inference problem, a Bayesian problem, a regression problem or a machine learning problem

maartenvsmeden.bsky.social
In medicine they are called "risk factors" and, of course, you want all "important" risk factors in your model all the time

Unless a risk factor is not statistically significant then you can drop that factor without issues

maartenvsmeden.bsky.social
Also, the fact that a model with the best AUC doesn't always mean the model makes the best predictions is lost in such cases too

maartenvsmeden.bsky.social
Surprisingly common thing: comparisons of prediction models developed using, say, Logistic Regression, Random Forest and XGBoost with conclusion XGBoost is "good" because it yields slightly higher AUC than LR or RF using the same data

Fact that "better" doesn't always mean "good" seems lost

Reposted by Maarten van Smeden

richarddriley.bsky.social
* New preprint led by Joao Matos & @gscollins.bsky.social

"Critical Appraisal of Fairness Metrics in Clinical Predictive AI"

- Important, rapidly growing area
- But confusion exists
- 62 fairness metrics identified so far
- Better standards & metrics needed for healthcare
arxiv.org/abs/2506.17035