Hilde Kuehne
@hildekuehne.bsky.social
2.4K followers 1.1K following 500 posts
Professor for CS at the Tuebingen AI Center and affiliated Professor at MIT-IBM Watson AI lab - Multimodal learning and video understanding - GC for ICCV 2025 - https://hildekuehne.github.io/
Posts Media Videos Starter Packs
hildekuehne.bsky.social
Nice… need to check if the representation learning community found a similar signal…
Reposted by Hilde Kuehne
ekazakos.bsky.social
To keep the tradition, the lineup is 🔥🔥🔥
gtolias.bsky.social
The Visual Recognition Group at CTU in Prague organizes the 50th Pattern Recognition and Computer Vision Colloquium with
Torsten Sattler, Paul-Edouard Sarlin, Vicky Kalogeiton, Spyros Gidaris, Anna Kukleva, and Lukas Neumann.
On Thursday Oct 9, 11:00-17:00.

cmp.felk.cvut.cz/colloquium/
hildekuehne.bsky.social
is there a paper I missed?
Reposted by Hilde Kuehne
gtolias.bsky.social
The Visual Recognition Group at CTU in Prague organizes the 50th Pattern Recognition and Computer Vision Colloquium with
Torsten Sattler, Paul-Edouard Sarlin, Vicky Kalogeiton, Spyros Gidaris, Anna Kukleva, and Lukas Neumann.
On Thursday Oct 9, 11:00-17:00.

cmp.felk.cvut.cz/colloquium/
hildekuehne.bsky.social
It looks so good… 😋… I‘m trying more veggies, no sweets but I‘m failing all the time
Reposted by Hilde Kuehne
dlcountdown.bsky.social
CVPR'26 (abs): 33 days.
CVPR'26 (paper): 40 days.
ECCV'26: 152 days.
Reposted by Hilde Kuehne
serge.belongie.com
@belongielab.org welcomes applications for a PhD position in CV/ML (fine-grained analysis of multimodal data, 2D/3D generative models, misinformation detection, self-supervised learning) / Apply though the ELLIS portal / Deadline 31-Oct-2025
Reposted by Hilde Kuehne
iccv.bsky.social
If you have an accepted paper at the main #ICCV2025 conference, please upload a thumbnail, a full poster image, and a 5 minute YouTube video presentation for your paper.

Please have the uploads done by October 8, 2025 aoe.

Instructions: iccv.thecvf.com/Conferences/...
Reposted by Hilde Kuehne
iccv.bsky.social
Hi all! Reminder that the #ICCV2025 Main Conference Poster Art And Video Upload Deadline is coming up on Oct 09 '25 04:00 AM UTC

Clock is ticking for the conference itself as well and we are very excited to see you all there 🥳
Reposted by Hilde Kuehne
dlcountdown.bsky.social
CVPR'26 (abs): 35 days.
CVPR'26 (paper): 42 days.
ECCV'26: 154 days.
hildekuehne.bsky.social
But this is not really something which is quantitatively shown in the paper …
hildekuehne.bsky.social
No comment on the second one ... but there should be a rule that all flavours of "emerging" or "emergence" are only allowed in context of models where the complete training data (I mean for the full pipeline) is public.

Everything else should be considered false marketing until proven otherwise.
a cartoon of a girl holding a bag and pointing at a woman
ALT: a cartoon of a girl holding a bag and pointing at a woman
media.tenor.com
hildekuehne.bsky.social
But what we know is that data is full of those basic samples ... just crawl enough conference talks and I show you emergence. And going from low to high res? ... never seen a video like that! And we didn't even touch style transfer.
So nope! But please, help me find the substance here. 4/4
hildekuehne.bsky.social
Second, we know there is tons of data and task leakage. So, without knowing the training data (come on everybody, sharing is caring!), we can not say anything about new "emerge whatever" capabilities (which have been rarely shown for real up to this point).
3/4
hildekuehne.bsky.social
Can you find any serious evaluation in this paper on any standard benchmark and metric for any task they claim? The main evaluation is "We run it ten times and test if it works in one of those cases". Would you accept this evaluation from any computer vision MSc student? Or run a car on that? 2/4
hildekuehne.bsky.social
omg ... where should I start, where should I end ... that rant would go on forever. Probably, to get a glimpse, read it as a research paper ... I mean, at least it was published on arxiv, so we should treat it with minimum respect. Let's go ... 1/4

And please don't take the rant too serious...
Reposted by Hilde Kuehne
paulgavrikov.bsky.social
🚨 New paper out!
"VisualOverload: Probing Visual Understanding of VLMs in Really Dense Scenes"
👉 arxiv.org/abs/2509.25339
We test 37 VLMs on 2,700+ VQA questions about dense scenes.
Findings: even top models fumble badly—<20% on the hardest split and key failure modes in counting, OCR & consistency.
Reposted by Hilde Kuehne
dlcountdown.bsky.social
CVPR'26 (abs): 37 days.
CVPR'26 (paper): 44 days.
ECCV'26: 156 days.
Reposted by Hilde Kuehne
ml4science.bsky.social
We're back at Neckawa in Tübingen today for Day 2 of our Cluster Conference #MLinScience2025! Our first speaker is @hildekuehne.bsky.social with her talk "Advances in Self-Supervised Multimodal Learning". Today's program: uni-tuebingen.de/en/165313#c2...
Scientist Hilde Kühne stands on a stage and gives a talk. On the wall behind her are slides with images illustrating her talk.
hildekuehne.bsky.social
Strands #571
“Simply divine”
🔵🔵🔵🟡
🔵🔵
hildekuehne.bsky.social
Connections
Puzzle #837
🟨🟨🟨🟨
🟩🟩🟩🟩
🟦🟦🟦🟦
🟪🟪🟪🟪
Reposted by Hilde Kuehne
csprofkgd.bsky.social
At #ICCV2025 we’re hosting the 1st workshop on Geometry-Free Novel View Synthesis & Controllable Video Models! (Or simply: “3D Vision in the era of Video Models”)

Our speaker lineup is 🔥🔥🔥 Better grab a seat early 😉

geofreenvs.github.io
Reposted by Hilde Kuehne
davidpicard.bsky.social
I'm currently attending the video Symposium sites.google.com/view/video-a...
Starting with already impressively good talks 🤩