toshi2k2.bsky.social
@toshi2k2.bsky.social
This is a massive piece of the puzzle for Explainable AI and Mechanistic Interpretability. 🧩

Researchers have hypothesized "Universality" - that different models learn similar circuits to solve tasks.

Our work provides the strongest proof, till date, for this hypothesis, and beyond features.
December 4, 2025 at 9:09 PM
The Universal Subspace unlocks the trifecta of Efficient AI:

1️⃣ Model Merging: Combine models analytically without hyperparameter tuning.
2️⃣ Efficient Training: Adapt to new tasks by learning tiny coefficients, not full weights.
3️⃣ Compression: 100x+ memory reduction.
December 4, 2025 at 9:05 PM
The Evidence: Look at the spectral decay (scree plots).

Whether it’s Mistral LoRAs or Vision Transformers, the variance drops off a cliff. A tiny handful of 'unique' directions capture nearly ALL the information.

It’s not random. It’s a signature.
December 4, 2025 at 9:02 PM
The "Independence" Myth: We usually assume that a model trained on medical X-rays lives in a different parameter space than one trained on satellite imagery.

We found the opposite.

Using spectral decomposition, we saw that weights from disjoint tasks converge to the same principal directions.
December 4, 2025 at 8:58 PM