majhas.bsky.social
@majhas.bsky.social
PhD Student at Mila & University of Montreal | Generative modeling, sampling, molecules
majhas.github.io
The gifs didn't post properly πŸ˜…

Here is one showing the electron cloud in two stages: (1) the learning of electron density during training and (2) the predicted ground-state across conformations 😎
June 10, 2025 at 10:06 PM
(9/9)⚑ Runtime efficiency
Self-refining training reduces total runtime up to 4 times compared to the baseline
and up to 2 times compared to the fully-supervised approach!!!
Less need for large pre-generated datasets β€” training and sampling happen in parallel.
June 10, 2025 at 7:49 PM
(8/n) πŸ§ͺ Robust generalization
We simulate molecular dynamics using each model’s energy predictions and evaluate accuracy along the trajectory.
Models trained with self-refinement stay accurate even far from the training distribution β€” while baselines quickly degrade.
June 10, 2025 at 7:49 PM
(7/n) πŸ“Š Performance under data scarcity
Our method achieves low energy error with as few as 25 conformations.
With 10Γ— less data, it matches or outperforms fully supervised baselines.
This is especially important in settings where labeled data is expensive or unavailable.
June 10, 2025 at 7:49 PM
(6/n) This minimization leads to Self-Refining Training:
πŸ” Use the current model to sample conformations via MCMC
πŸ“‰ Use those conformations to minimize energy and update the model

Everything runs asynchronously, without need for labeled data and minimal number of conformations from a dataset!
June 10, 2025 at 7:49 PM
(5/n) To get around this, we introduce a variational upper bound on the KL between any sampling distribution q(R) and the target Boltzmann distribution.

Jointly minimizing this bound wrt ΞΈ and q yields
βœ… A model that predicts the ground-state solutions
βœ… Samples that match the ground true density
June 10, 2025 at 7:49 PM
(4/n) With an amortized DFT model f_ΞΈ(R), we define the density of molecular conformations as the
Boltzmann distribution

This isn't a typical ML setup because
❌ No samples from the density - can’t train a generative model
❌ No density - can’t sample via Monte Carlo!
June 10, 2025 at 7:49 PM
(3/n) DFT offers a scalable solution to the SchrΓΆdinger equation but must be solved independently for each geometry by minimizing energy wrt coefficients C for a fixed basis.

This presents a bottleneck for MD/sampling.

We want to amortize this - train a model that generalizes across geometries R.
June 10, 2025 at 7:49 PM
(2/n) This work is the result of an amazing collaboration with @fntwin.bsky.social Hatem Helal @dom-beaini.bsky.social @k-neklyudov.bsky.social
GitHub - majhas/self-refining-dft
Contribute to majhas/self-refining-dft development by creating an account on GitHub.
github.com
June 10, 2025 at 7:49 PM
(1/n)🚨Train a model solving DFT for any geometry with almost no training data
Introducing Self-Refining Training for Amortized DFT: a variational method that predicts ground-state solutions across geometries and generates its own training data!
πŸ“œ arxiv.org/abs/2506.01225
πŸ’» github.com/majhas/self-...
June 10, 2025 at 7:49 PM
Reposted
New preprint! πŸ§ πŸ€–

How do we build neural decoders that are:
⚑️ fast enough for real-time use
🎯 accurate across diverse tasks
🌍 generalizable to new sessions, subjects, and even species?

We present POSSM, a hybrid SSM architecture that optimizes for all three of these axes!

🧡1/7
June 6, 2025 at 5:40 PM
Reposted
🧡(1/7) Have you ever wanted to combine different pre-trained diffusion models but don't have time or data to retrain a new, bigger model?

πŸš€ Introducing SuperDiff πŸ¦Ήβ€β™€οΈ – a principled method for efficiently combining multiple pre-trained diffusion models solely during inference!
December 28, 2024 at 2:32 PM
Reposted
πŸ”Š Super excited to announce the first ever Frontiers of Probabilistic Inference: Learning meets Sampling workshop at #ICLR2025 @iclr-conf.bsky.social!

πŸ”— website: sites.google.com/view/fpiwork...

πŸ”₯ Call for papers: sites.google.com/view/fpiwork...

more details in thread belowπŸ‘‡ 🧡
December 18, 2024 at 7:09 PM
Reposted
Now you can generate equilibrium conformations for your small molecule in 3 lines of code with ET-Flow! Awesome effort put in by @fntwin.bsky.social!
December 12, 2024 at 4:37 PM
Reposted
ET-Flow shows, once again, that equivariance is better than Transformer when physical precision matters!

come see us at @neuripsconf.bsky.social !!
December 7, 2024 at 3:57 PM
Excited to share our work! I had a wonderful time collaborating with these brilliant people
We’re excited to present ET-Flow at #NeurIPS 2024β€”an Equivariant Flow Matching model that combines simplicity, efficiency, and precision to set a new standard for 3D molecular conformer generation.
πŸ”–Paper: arxiv.org/abs/2410.22388
πŸ”—Github: github.com/shenoynikhil...
December 7, 2024 at 4:01 PM