ExplainableML
@eml-munich.bsky.social
470 followers 24 following 120 posts
Institute for Explainable Machine Learning at @www.helmholtz-munich.de and Interpretable and Reliable Machine Learning group at Technical University of Munich and part of @munichcenterml.bsky.social
Posts Media Videos Starter Packs
eml-munich.bsky.social
During his PhD, Karsten interned at Mata AI and Googlem DeepMind, working on generalization in representation learning and large-scale multimodal pretraining techniques.

👇Checkout his selected publications in top-tier conferences such as NeurIPS, ICLR, CVPR or ICCV:
eml-munich.bsky.social
🦾 (Multimodal) model pretraining
🧠 Model generalization, reuse and transferability research
🎆 Continual (multimodal) training of such models.
eml-munich.bsky.social
Karsten has been an ELLIS and IMPRS-IS PhD student since May 2021, supervised by both @zeynepakata.bsky.social and Oriol Vinyals. His research has been centered around robust and effective deployment of (large) neural networks in the real world, with particular focus on:
eml-munich.bsky.social
🎓PhD Spotlight: Karsten Roth

Celebrate @confusezius.bsky.social , who defended his PhD on June 24th summa cum laude!

🏁 His next stop: Google DeepMind in Zurich!

Join us in celebrating Karsten's achievements and wishing him the best for his future endeavors! 🥳
Reposted by ExplainableML
simonroschmann.bsky.social
This project was a collaboration between @eml-munich.bsky.social and Huawei Paris Noah’s Ark Lab. Thank you to my collaborators @qbouniot.bsky.social, Vasilii Feofanov, Ievgen Redko, and particularly to my advisor @zeynepakata.bsky.social for guiding me through my first PhD project!
eml-munich.bsky.social
5/
BayesCap: Bayesian Identity Cap for Calibrated Uncertainty in Frozen Neural Networks (ECCV 2022)
Uddeshya Upadhyay* , @shyamgopal.bsky.social * , Yanbei Chen, Massimiliano Mancini, and @zeynepakata.bsky.social
[Paper]: arxiv.org/pdf/2207.06873
[Code]: github.com/ExplainableM...
arxiv.org
eml-munich.bsky.social
4/
KG-SP: Knowledge Guided Simple Primitives for Open World Compositional Zero-Shot Learning (CVPR 2022)
@shyamgopal.bsky.social , Massimiliano Mancini, @zeynepakata.bsky.social
[Paper]: arxiv.org/pdf/2205.06784
[Code]: github.com/ExplainableM...
arxiv.org
eml-munich.bsky.social
3/
Vision-by-Language for Training-Free Compositional Image Retrieval (ICLR 2024)
@shyamgopal.bsky.social *, @confusezius.bsky.social *, Massimiliano Mancini, @zeynepakata.bsky.social
[Paper]: arxiv.org/pdf/2310.09291
[Code]: github.com/ExplainableM...
arxiv.org
eml-munich.bsky.social
Join us in celebrating Shyam’s achievements and wishing him the best for his future endeavors! 🥳

👇Checkout his selected publications in top-tier conferences such as CVPR, ICCV, ECCV, NeurIPS, ICLR:
eml-munich.bsky.social
More recently, he interned at Snap Inc. , where he worked on enhancing diffusion models through preference tuning.
eml-munich.bsky.social
His research has been centred around compositionality in vision and language. In particular, he has focused on:

🧠 Inference-time alignment and preference optimization for multimodal models

🏁 Composed retrieval, text-to-image generation, and uncertainty estimation
eml-munich.bsky.social
🎓PhD Spotlight: Shyamgopal Karthik

Celebrate @shyamgopal.bsky.social , who will defend his PhD on 23rd June! Shyam has been a PhD student @unituebingen.bsky.social since October 2021, supervised by @zeynepakata.bsky.social.
Reposted by ExplainableML
sukrutrao.bsky.social
Join us in taking stock of the state of the field of explainability in computer vision, at our Workshop on Explainable Computer Vision: Quo Vadis? at #ICCV2025!

@iccv.bsky.social
Call for papers at the eXCV workshop at ICCV 2025.
eml-munich.bsky.social
(4/4) FLAIR: VLM with Fine-grained Language-informed Image Representations
@rui-xiao.bsky.social will present his work on pretraining a CLIP-like model that generates fine-grained image representations.
📍 ExHall D Poster #368
⏲️ Sun 15 Jun 10:30 a.m. CDT — 12:30 p.m. CDT
eml-munich.bsky.social
(3/4) How to Merge Your Multimodal Models Over Time?
@confusezius.bsky.social will also present this amazing work that introduces a unified framework for temporal model merging.
📍 ExHall D Poster #445
⏲️ Sat 14 Jun 5 p.m. CDT — 7 p.m. CDT
eml-munich.bsky.social
(2/4) COSMOS: Cross-Modality Self-Distillation for Vision Language Pre-training
Sanghwan Kim will present COSMOS, integrating a novel text-cropping strategy and cross-attention module into a self-supervised learning framework.
📍 ExHall D Poster #387
⏲️ Sat 14 Jun 10:30 a.m. CDT — 12:30 p.m. CDT