Giorgos Kordopatis-Zilos
@gkordo.bsky.social
110 followers 270 following 24 posts
Postdoct Researcher at Visual Recognition Group, CTU in Prague - gkordo.github.io/
Posts Media Videos Starter Packs
gkordo.bsky.social
• SigLIP2 → still the best for text-to-image retrieval.
– Unlike SigLIP models, PE shows a large gap between its image-to-image and text-to-image performance.
gkordo.bsky.social
• PE → achieves state-of-the-art in image-to-image retrieval with linear adaptation.
  – Gains +12% performance from adaptation.
  – Outperforms SigLIP2 (previous best) by +2%.
  – DINOv3 lags behind, gaining less than +2% from adaptation
gkordo.bsky.social
🔍 Key results from our new evaluation on ILIAS:

• DINOv3 → sets a new state-of-the-art in image-to-image retrieval without linear adaptation
  – The large variant outperforms all other models by a significant margin.
  – The base outperforms the large variants of other model series.
gkordo.bsky.social
🚀 new state-of-the-art on ILIAS dataset!

Curious how well the latest models can recognize particular objects?
We evaluated the base and large variants of DINOv3 and Perception Encoder (PE) on instance-level image retrieval.

See the results 👉 vrg.fel.cvut.cz/ilias/
Reposted by Giorgos Kordopatis-Zilos
gtolias.bsky.social
The Colloquium in Pattern Recognition and Computer Vision of the Visual Recognition Group at CTU in Prague has a long tradition dating back to 1998. The list of all speakers is available docs.google.com/spreadsheets.... Enjoy! The 50th edition is coming soon cmp.felk.cvut.cz/colloquium/
Pattern Recognition and Computer Vision Colloquium - past speakers
docs.google.com
Reposted by Giorgos Kordopatis-Zilos
ducha-aiki.bsky.social
Processing and acquisition traces in visual encoders: What does CLIP know about your camera?

@ryan-ramos.bsky.social @stojnicv.xyz @gkordo.bsky.social Yuta Nakashima @gtolias.bsky.social
@noagarciad.bsky.social
tl;dr: CLIP sees difference DSLR vs iphone, DINO doesn't.
arxiv.org/abs/2508.10637
1/
Reposted by Giorgos Kordopatis-Zilos
stojnicv.xyz
Have you ever asked yourself how much your favorite vision model knows about image capture parameters (e.g., the amount of JPEG compression, the camera model, etc.)? Furthermore, could these parameters influence its semantic recognition abilities?
gkordo.bsky.social
Topics:
- IL object classification, detectection, segmentation, and pose estimation
- particular object and event retrieval
- personalized image/video generation
- cross/multi-modal recognition at IL
- image matching, place recognition, video tracking
- other ILR+G applications
- ILR+G datasets
gkordo.bsky.social
We invite papers on topics related to instance-level recognition and/or generation. We extend the submission deadline to give the authors of related papers that didn't make it to ICCV 2025 a chance to still join us in Hawaii 🌺
gkordo.bsky.social
🚨 Deadline Extension

Instance-Level Recognition and Generation (ILR+G) Workshop at ICCV2025 @iccv.bsky.social

📅 new deadline: June 26, 2025 (23:59 AoE)
📄 paper submission: cmt3.research.microsoft.com/ILRnG2025
🌐 ILR+G website: ilr-workshop.github.io/ICCVW2025/

#ICCV2025 #ComputerVision #AI
Reposted by Giorgos Kordopatis-Zilos
spyrosgidaris.bsky.social
I am at #CVPR2025 this week in Nashville!

Presenting "Advancing Semantic Future Prediction through Multimodal Visual Sequence Transformers" on multi-modal semantic future prediction.

Come discuss!

Fri 13 Jun 10:30-12:30, poster #345
bsky.app/profile/sta8...
sta8is.bsky.social
🧵 Excited to share our latest work: FUTURIST - A unified transformer architecture for multimodal semantic future prediction, is accepted to #CVPR2025! Here's how it works (1/n)
👇 Links to the arxiv and github below
Reposted by Giorgos Kordopatis-Zilos
stojnicv.xyz
Are you at @cvprconference.bsky.social #CVPR2025 ? Come and check out LPOSS.

We show how can graph-based label propagation be used to improve weak, patch-level predictions from VLMs for open-vocabulary semantic segmentation.

📅 June 13, 2025, 16:00 – 18:00 CDT
📍 Location: ExHall D, Poster #421
Reposted by Giorgos Kordopatis-Zilos
gtolias.bsky.social
VRG is presenting 8 papers at #CVPR2025. You can find me and collaborators at the following 4 posters:

Fri 10:30-12:30 A Dataset for Semantic Segmentation in the Presence of Unknowns
Fri 16:00-18:00 LOCORE: Image Re-ranking with Long-Context Sequence Modeling
gkordo.bsky.social
Are you at @cvprconference.bsky.social? Come by our poster!
📅 Sat 14/6, 10:30-12:30
📍 Poster #395, ExHall D
Reposted by Giorgos Kordopatis-Zilos
noagarciad.bsky.social
The 2nd CEGIS workshop on visual generative models evaluation is back at #ICCV2025!!

Submit your contributions:
- Deadline: June 26th 2025
- Notification: July 10th, 2025
- Camera-ready: August 18th, 2025

See you in Honolulu!

sites.google.com/view/cegis-w...

@iccv.bsky.social
cegis
2nd workshop on critical evaluation of generative models and their impact on society 19 or 20 October 2025 at ICCV 2025, Honolulu, Hawaii
sites.google.com
gkordo.bsky.social
Call for Papers update - ILR+G workshop @iccv.bsky.social

We will now feature a single submission track with new submission dates.

📅 New submission deadline: June 21, 2025
🔗 Submit here: cmt3.research.microsoft.com/ILRnG2025
🌐 More details: ilr-workshop.github.io/ICCVW2025/

#ICCV2025
Reposted by Giorgos Kordopatis-Zilos
gtolias.bsky.social
VRG from CTU in Prague has 9 of its members listed as outstanding reviewers. Congratulations to @gkordo.bsky.social, @billpsomas.bsky.social , @stojnicv.xyz , Pavel Suma, @ducha-aiki.bsky.social , Miroslav Purkrábek, Tomas Vojir, and Yaqing Ding.
cvprconference.bsky.social
Behind every great conference is a team of dedicated reviewers. Congratulations to this year’s #CVPR2025 Outstanding Reviewers!

cvpr.thecvf.com/Conferences/...
gkordo.bsky.social
🚨 Call for Papers!

7th Instance-Level Recognition and Generation (ILR+G) Workshop at @iccv.bsky.social

📍 Honolulu, Hawaii 🌺
📅 October 19–20, 2025
🌐 ilr-workshop.github.io/ICCVW2025/

in-proceedings deadline: June 7
out-of-proceedings deadline: June 30

#ICCV2025
ILR+G 2025
The Official Site of ICCV 2025 Workshop, Instance-Level Recognition and Generation Workshop
ilr-workshop.github.io
gkordo.bsky.social
also, thanks to @nielsrogge.bsky.social for the initial contact for hosting ILIAS on @hf.co
gkordo.bsky.social
Exciting News!
Shoutout to @hf.co 🤗 for hosting ILIAS in its full scale!

You can now easily download it via:
huggingface.co/datasets/vrg...

This marks a milestone in i2i and t2i retrieval, paving the way for significant advancements in the field.

Huge thanks to @hf.co team for their support!
gkordo.bsky.social
yes, we plan to release the top-1k shortlist of SigLIP, which will be used in the evaluation code. But we can provide it for any benchmarked model if useful for someone.
gkordo.bsky.social
correct, it is not that VLMs perform well. Much better compared with older models, but still performance is very low. SigLIP mAP@10=0.28, so no significant difference.
gkordo.bsky.social
Yeap, until the introduction of VLMs, global representation models would really struggle to perform well. But they are steadily getting better the last years.