🇪🇺 PhD student at the Max Planck Institute for Informatics, and Institute of Science & Technology - Austria.
💻🏃🏻♂️🚴🏻🏋🏻🏊⛷️🎸🎹📚
Webpage: https://sidgairo18.github.io/
This is joint work with Adam Wróbel (project lead), Jacek Tabor, Bernt Schiele, Bartosz Zieliński, and Dawid Rymarczyk.
📰Preprint: arxiv.org/abs/2602.06613
💻Code: github.com/a-vrobell/DAVE [to be released soon]
This is joint work with Adam Wróbel (project lead), Jacek Tabor, Bernt Schiele, Bartosz Zieliński, and Dawid Rymarczyk.
📰Preprint: arxiv.org/abs/2602.06613
💻Code: github.com/a-vrobell/DAVE [to be released soon]
Vs prior methods, DAVE gives sharper, more object-centric attributions with less background + fewer patch-grid artifacts.
Vs prior methods, DAVE gives sharper, more object-centric attributions with less background + fewer patch-grid artifacts.
On inherently interpretable B-cos ViTs, DAVE yields sharper, more object-aligned maps and improves localization vs built-in B-cos explanations.
B-cos ViTs produce sharp attributions only when trained with a conv-stem, DAVE fixes this reliance.
On inherently interpretable B-cos ViTs, DAVE yields sharper, more object-aligned maps and improves localization vs built-in B-cos explanations.
B-cos ViTs produce sharp attributions only when trained with a conv-stem, DAVE fixes this reliance.
Consistent improvements on localization metrics and perturbation evaluations.
Consistent improvements on localization metrics and perturbation evaluations.
Sample small spatial transforms + noise, compute effective transformation (conditioned forward blocks gradients through conditioning), inverse-transform & average, then apply element-wise to the input.
Sample small spatial transforms + noise, compute effective transformation (conditioned forward blocks gradients through conditioning), inverse-transform & average, then apply element-wise to the input.
DAVE adds low-pass stabilization by averaging the equivariant effective transformation under small input perturbations (Gaussian smoothing in expectation).
This removes components unstable to tiny input changes. (See above fig., last column.)
DAVE adds low-pass stabilization by averaging the equivariant effective transformation under small input perturbations (Gaussian smoothing in expectation).
This removes components unstable to tiny input changes. (See above fig., last column.)
Even the effective transformation can carry architecture-induced grid patterns.
DAVE filters them by enforcing local equivariance: under small spatial transforms, the attribution must transform consistently.
Even the effective transformation can carry architecture-induced grid patterns.
DAVE filters them by enforcing local equivariance: under small spatial transforms, the attribution must transform consistently.
Operator variation can amplify tiny perturbations → high-frequency junk in attributions.
DAVE drops this term and keeps the effective transformation as a cleaner attribution operator.
Operator variation can amplify tiny perturbations → high-frequency junk in attributions.
DAVE drops this term and keeps the effective transformation as a cleaner attribution operator.
Model each ViT layer as an input-dependent linear operator L(X) applied to X.
Then the input-gradient decomposes into:
(1) effective transformation L(X)
(2) operator variation (how L changes w.r.t. X)
Model each ViT layer as an input-dependent linear operator L(X) applied to X.
Then the input-gradient decomposes into:
(1) effective transformation L(X)
(2) operator variation (how L changes w.r.t. X)
ViT components (patch embedding, attention routing, etc.) inject structured artifacts into gradients → explanations become noisy/unstable, or methods fall back to coarse patch-level maps.
ViT components (patch embedding, attention routing, etc.) inject structured artifacts into gradients → explanations become noisy/unstable, or methods fall back to coarse patch-level maps.
📚Feel free to share, bookmark, and contribute here (github.com/sidgairo18/s...)!
P.S.: Please feel free to share relevant resources in the comments / thread and I'll add them as well 😀 (n/n)
📚Feel free to share, bookmark, and contribute here (github.com/sidgairo18/s...)!
P.S.: Please feel free to share relevant resources in the comments / thread and I'll add them as well 😀 (n/n)
A practical checklist-driven guide on writing with clarity, rigor, and reproducibility. Inspired by ICML's best practices and more.
🔗https://sidgairo18.github.io/how_to_write_academic_papers.html (4/n)
A practical checklist-driven guide on writing with clarity, rigor, and reproducibility. Inspired by ICML's best practices and more.
🔗https://sidgairo18.github.io/how_to_write_academic_papers.html (4/n)
What makes a good review? This guide compiles best practices from ICML, ICLR, CVPR, and other leading conferences.
🔗https://sidgairo18.github.io/how_to_review_scientific_papers.html (3/n)
What makes a good review? This guide compiles best practices from ICML, ICLR, CVPR, and other leading conferences.
🔗https://sidgairo18.github.io/how_to_review_scientific_papers.html (3/n)
Mindset, habits, tools, writing, productivity, and advice.
🔗https://sidgairo18.github.io/notes_and_resources_on_how_to_do_research.html (2/n)
Mindset, habits, tools, writing, productivity, and advice.
🔗https://sidgairo18.github.io/notes_and_resources_on_how_to_do_research.html (2/n)