hochadel.bsky.social
hochadel.bsky.social
@hochadel.bsky.social
Reposted by hochadel.bsky.social
w h a t

"Pruning as few as a single parameter can destroy an LLM's ability to generate text -- increasing perplexity by 3 orders of magnitude and reducing zero-shot accuracy to guessing."
The Super Weight in Large Language Models
Recent works have shown a surprising result: a small fraction of Large Language Model (LLM) parameter outliers are disproportionately important to the quality of the model. LLMs contain billions of pa...
arxiv.org
May 27, 2025 at 9:11 PM