Lars van der Laan
@larsvanderlaan3.bsky.social
570 followers 120 following 24 posts
Ph.D. Student @uwstat; Research fellowship @Netflix; visiting researcher @UCJointCPH; M.A. @UCBStatistics - machine learning; calibration; semiparametrics; causal inference. https://larsvanderlaan.github.io
Posts Media Videos Starter Packs
Reposted by Lars van der Laan
alxndrmlk.bsky.social
He did it before Double Machine Learning

I met with professor Mark van der Laan because I think his work is pretty incredible and it sometimes feels like a secret that only a few people know about, especially in industry.

1/

#CausalSky #StatSky #CausalInference
larsvanderlaan3.bsky.social
What does ‘biased’ mean here? It would be biased in expectation, since if you were to repeat the experiment many times, some power users would join. If you define your estimator as the empirical mean over non-power users, then it might be unbiased.
larsvanderlaan3.bsky.social
I’d be surprised if this actually works in practice, since neural networks are often overfitting (e.g. perfectly fitting labels with double descent), which violates donsker conditions. And, the neural tangent kernel ridge approximation of neural networks has been shown to not hold empirically.
Reposted by Lars van der Laan
alexluedtke.bsky.social
I've advised 15 PhD students—10 were international students. All graduates continue advancing U.S. excellence in research and education. Cutting off this pipeline of talent would be shortsighted.
Reposted by Lars van der Laan
apoorvalal.com
I had a hard time believing it was as simple as this until Lars taught me how to implement it - calibrate=True and you're done

github.com/apoorvalal/a...
larsvanderlaan3.bsky.social
Calibrate your outcome predictions and propensities using isotonic regression as follows:

mu_hat <- as.stepfun(isoreg(mu_hat, Y))(mu_hat)

pi_hat <- as.stepfun(isoreg(pi_hat, A))(pi_hat)

(Or use the isoreg_with_xgboost function given in the paper, which I recommend)
larsvanderlaan3.bsky.social
Had a great time presenting at #ACIC on doubly robust inference via calibration

Calibrating nuisance estimates in DML protects against model misspecification and slow convergence.

Just one line of code is all it takes.
Reposted by Lars van der Laan
paperposterbot.bsky.social
link 📈🤖
Nonparametric Instrumental Variable Inference with Many Weak Instruments (Laan, Kallus, Bibaut) We study inference on linear functionals in the nonparametric instrumental variable (NPIV) problem with a discretely-valued instrument under a many-weak-instruments asymptotic regime, where the
Reposted by Lars van der Laan
larsvanderlaan3.bsky.social
I’ll be giving an oral presentation at ACIC in the Advancing Causal Inference session with ML on Wednesday!

My talk will be on Automatic Double Reinforcement Learning and long term causal inference!

I’ll discuss Markov decision processes, Q-functions, and a new form of calibration for RL!
larsvanderlaan3.bsky.social
New preprint with #Netflix out!

We study the NPIV problem with a discrete instrument under a many-weak-instruments regime.

A key application: constructing confounding-robust surrogates using past experiments as instruments.

My mentor Aurélien Bibaut will be presenting a poster at #ACIC2025!
arxiv-stat-me.bsky.social
Lars van der Laan, Nathan Kallus, Aur\'elien Bibaut
Nonparametric Instrumental Variable Inference with Many Weak Instruments
https://arxiv.org/abs/2505.07729
larsvanderlaan3.bsky.social
Our work on stabilized inverse probability weighting via calibration was accepted to #CLeaR2025! I gave an oral presentation last week and was honored to receive the Best Paper Award.

I’ll be giving a related poster talk at #ACIC on calibration and DML and how it provides doubly robust inference!
arxiv-stat-me.bsky.social
Lars van der Laan, Ziming Lin, Marco Carone, Alex Luedtke
Stabilized Inverse Probability Weighting via Isotonic Calibration
https://arxiv.org/abs/2411.06342
larsvanderlaan3.bsky.social
This work is a result of my internship at Netflix over the summer and is joint with Aurelien Bibaut and Nathan Kallus.
larsvanderlaan3.bsky.social
I’ll be giving an oral presentation at ACIC in the Advancing Causal Inference session with ML on Wednesday!

My talk will be on Automatic Double Reinforcement Learning and long term causal inference!

I’ll discuss Markov decision processes, Q-functions, and a new form of calibration for RL!
larsvanderlaan3.bsky.social
Inference for smooth functionals of M-estimands in survival models, like regularized coxPH and the beta-geometric model (see our experiments section) are one application of this approach.
larsvanderlaan3.bsky.social
By targeting low dimensional summaries, there is no need to establish asymptotic normality of the entire infinite dimensional M-estimator (which isn’t possible in general). It allows for the use of ML and regularization to estimate it, and valid inference via a one step bias correction.
larsvanderlaan3.bsky.social
If you’re willing to consider smooth functionals of the infinite dimensional M-estimand, then there is a general theory for inference, where the sandwich variance estimator now involves the derivative of the loss and a Riesz representer of the functional.

Working paper:
arxiv.org/pdf/2501.11868
arxiv.org
larsvanderlaan3.bsky.social
The motivation should have been something like a confounder that is somewhat predictive of both the treatment and outcome might be more important to adjust for then a variable that is super predictive of the outcome but doesn’t predict treatment. TR might help give more importance to such variables
larsvanderlaan3.bsky.social
One could have given an analogous theorem saying that E[Y | T, X] is a sufficient deconfounding score and argued that one should only adjust for features predictive of the outcome. So yeah I think it’s wrong/poorly phrased
larsvanderlaan3.bsky.social
The OP’s approach is based on the conditional probability of Y given the treatment is intervened upon and set to some value. But, they don’t seem to define what this means formally, which is exactly what potential outcomes/NPSEM achieve.
larsvanderlaan3.bsky.social
The second stage coefficients are the estimand (identifying the structural coefficients/treatment effect). The first stage coefficients are nuisances, and typically not of direct interest.
larsvanderlaan3.bsky.social
🚨 Excited about this new paper on Generalized Venn Calibration and conformal prediction!

We show that Venn and Venn-Abers can be extended to general losses, and that conformal prediction can be viewed as Venn multicalibration for the quantile loss!

#calibration #conformal
arxiv-stat-ml.bsky.social
Lars van der Laan, Ahmed Alaa
Generalized Venn and Venn-Abers Calibration with Applications in Conformal Prediction
https://arxiv.org/abs/2502.05676
Reposted by Lars van der Laan
arxiv-stat-ml.bsky.social
Lars van der Laan, Ahmed Alaa
Generalized Venn and Venn-Abers Calibration with Applications in Conformal Prediction
https://arxiv.org/abs/2502.05676