@majhas.bsky.social
120 followers 110 following 11 posts
PhD Student at Mila & University of Montreal | Generative modeling, sampling, molecules majhas.github.io
Posts Media Videos Starter Packs
majhas.bsky.social
The gifs didn't post properly 😅

Here is one showing the electron cloud in two stages: (1) the learning of electron density during training and (2) the predicted ground-state across conformations 😎
majhas.bsky.social
(9/9)⚡ Runtime efficiency
Self-refining training reduces total runtime up to 4 times compared to the baseline
and up to 2 times compared to the fully-supervised approach!!!
Less need for large pre-generated datasets — training and sampling happen in parallel.
majhas.bsky.social
(8/n) 🧪 Robust generalization
We simulate molecular dynamics using each model’s energy predictions and evaluate accuracy along the trajectory.
Models trained with self-refinement stay accurate even far from the training distribution — while baselines quickly degrade.
majhas.bsky.social
(7/n) 📊 Performance under data scarcity
Our method achieves low energy error with as few as 25 conformations.
With 10× less data, it matches or outperforms fully supervised baselines.
This is especially important in settings where labeled data is expensive or unavailable.
majhas.bsky.social
(6/n) This minimization leads to Self-Refining Training:
🔁 Use the current model to sample conformations via MCMC
📉 Use those conformations to minimize energy and update the model

Everything runs asynchronously, without need for labeled data and minimal number of conformations from a dataset!
majhas.bsky.social
(5/n) To get around this, we introduce a variational upper bound on the KL between any sampling distribution q(R) and the target Boltzmann distribution.

Jointly minimizing this bound wrt θ and q yields
✅ A model that predicts the ground-state solutions
✅ Samples that match the ground true density
majhas.bsky.social
(4/n) With an amortized DFT model f_θ(R), we define the density of molecular conformations as the
Boltzmann distribution

This isn't a typical ML setup because
❌ No samples from the density - can’t train a generative model
❌ No density - can’t sample via Monte Carlo!
majhas.bsky.social
(3/n) DFT offers a scalable solution to the Schrödinger equation but must be solved independently for each geometry by minimizing energy wrt coefficients C for a fixed basis.

This presents a bottleneck for MD/sampling.

We want to amortize this - train a model that generalizes across geometries R.
majhas.bsky.social
(1/n)🚨Train a model solving DFT for any geometry with almost no training data
Introducing Self-Refining Training for Amortized DFT: a variational method that predicts ground-state solutions across geometries and generates its own training data!
📜 arxiv.org/abs/2506.01225
💻 github.com/majhas/self-...
Reposted
averyryoo.bsky.social
New preprint! 🧠🤖

How do we build neural decoders that are:
⚡️ fast enough for real-time use
🎯 accurate across diverse tasks
🌍 generalizable to new sessions, subjects, and even species?

We present POSSM, a hybrid SSM architecture that optimizes for all three of these axes!

🧵1/7
Reposted
k-neklyudov.bsky.social
🧵(1/7) Have you ever wanted to combine different pre-trained diffusion models but don't have time or data to retrain a new, bigger model?

🚀 Introducing SuperDiff 🦹‍♀️ – a principled method for efficiently combining multiple pre-trained diffusion models solely during inference!
Reposted
joeybose.bsky.social
🔊 Super excited to announce the first ever Frontiers of Probabilistic Inference: Learning meets Sampling workshop at #ICLR2025 @iclr-conf.bsky.social!

🔗 website: sites.google.com/view/fpiwork...

🔥 Call for papers: sites.google.com/view/fpiwork...

more details in thread below👇 🧵
Reposted
nikhilshenoy.bsky.social
Now you can generate equilibrium conformations for your small molecule in 3 lines of code with ET-Flow! Awesome effort put in by @fntwin.bsky.social!
Reposted
dom-beaini.bsky.social
ET-Flow shows, once again, that equivariance is better than Transformer when physical precision matters!

come see us at @neuripsconf.bsky.social !!
majhas.bsky.social
Excited to share our work! I had a wonderful time collaborating with these brilliant people
jyoonlee.bsky.social
We’re excited to present ET-Flow at #NeurIPS 2024—an Equivariant Flow Matching model that combines simplicity, efficiency, and precision to set a new standard for 3D molecular conformer generation.
🔖Paper: arxiv.org/abs/2410.22388
🔗Github: github.com/shenoynikhil...