Stella Frank
@scfrank.bsky.social
85 followers 400 following 19 posts
Thinking about multimodal representations | Postdoc at UCPH/Pioneer Centre for AI (DK).
Posts Media Videos Starter Packs
Reposted by Stella Frank
christophheilig.bsky.social
1/8 🧵 GPT-5's storytelling problems reveal a deeper AI safety issue. I've been testing its creative writing capabilities, and the results are concerning - not just for literature, but for AI development more broadly. 🚨
Reposted by Stella Frank
bonan.bsky.social
My Lab at the University of Edinburgh🇬🇧 has funded PhD positions for this cycle!

We study the computational principles of how people learn, reason, and communicate.

It's a new lab, and you will be playing a big role in shaping its culture and foundations.

Spread the words!
Reposted by Stella Frank
nicolasdufour.bsky.social
🚀 DinoV3 just became the new go-to backbone for geoloc!
It outperforms CLIP-like models (SigLip2, finetuned StreetCLIP)… and that’s shocking 🤯
Why? CLIP models have an innate advantage — they literally learn place names + images. DinoV3 doesn’t.
Reposted by Stella Frank
delliott.bsky.social
📢I am hiring a Postdoc to work on post-training methods for low-resource languages. Apply by August 15 employment.ku.dk/faculty/?sho....
Let's talk at #ACL2025NLP in Vienna if you want to know more about the position and life in Denmark.
Postdoc in Natural Language Processing
employment.ku.dk
Reposted by Stella Frank
abeba.bsky.social
New paper hot off the press www.nature.com/articles/s41...

We analysed over 40,000 computer vision papers from CVPR (the longest standing CV conf) & associated patents tracing pathways from research to application. We found that 90% of papers & 86% of downstream patents power surveillance

1/
Computer-vision research powers surveillance technology - Nature
An analysis of research papers and citing patents indicates the extensive ties between computer-vision research and surveillance.
www.nature.com
Reposted by Stella Frank
onslies.bsky.social
"Researching and reflecting on the harms of AI is not itself harm reduction. It may even contribute to rationalizing, normalizing, and enabling harm. Critical reflection without appropriate action is thus quintessentially critical washing."
scfrank.bsky.social
Fallacy of the Day:
Calling two different things by the same name doesn't make them the same (jingle) and calling the same thing by different names doesn't make them different (jangle)
en.wikipedia.org/wiki/Jingle-...

(this is going to be so useful for reviewing)
Jingle-jangle fallacies - Wikipedia
en.wikipedia.org
scfrank.bsky.social
Sad not to be there in person but this work will also be presented at ACL in Vienna 2025 - see you there!
Des presenting at VisCon CVPR 2025
scfrank.bsky.social
📯 Best Paper Award at CVPR workshop on Visual concepts for our (@doneata.bsky.social + @delliott.bsky.social) paper on probing vision/lang/ vision+lang models for semantic norms!

TLDR: SSL vision models (swinV2, dinoV2) are surprisingly similar to LLM & VLMs even w/o lang 👀
arxiv.org/abs/2506.03994
Reposted by Stella Frank
srishtiy.bsky.social
I am excited to announce our latest work 🎉 "Cultural Evaluations of Vision-Language Models Have a Lot to Learn from Cultural Theory". We review recent works on culture in VLMs and argue for deeper grounding in cultural theory to enable more inclusive evaluations.

Paper 🔗: arxiv.org/pdf/2505.22793
Paper title "Cultural Evaluations of Vision-Language Models
Have a Lot to Learn from Cultural Theory"
Reposted by Stella Frank
mariaa.bsky.social
Check out our new paper led by @srishtiy.bsky.social and @nolauren.bsky.social! This work brings together computer vision, cultural theory, semiotics, and visual studies to provide new tools and perspectives for the study of ~culture~ in VLMs.
srishtiy.bsky.social
I am excited to announce our latest work 🎉 "Cultural Evaluations of Vision-Language Models Have a Lot to Learn from Cultural Theory". We review recent works on culture in VLMs and argue for deeper grounding in cultural theory to enable more inclusive evaluations.

Paper 🔗: arxiv.org/pdf/2505.22793
Paper title "Cultural Evaluations of Vision-Language Models
Have a Lot to Learn from Cultural Theory"
Reposted by Stella Frank
kyunghyuncho.bsky.social
as an extra take-away, this implies that our eval tends to be overly precision focused. we should really think of what we lose in terms of recalls, as this directly relates to what we miss out for whom when we build these large-scale, general-purpose models.

(4/4)
Reposted by Stella Frank
cohereforai.bsky.social
🚀 We are excited to introduce Kaleidoscope, the largest culturally-authentic exam benchmark.

📌 Most VLM benchmarks are English-centric or rely on translations—missing linguistic & cultural nuance. Kaleidoscope expands in-language multilingual 🌎 & multimodal 👀 VLMs evaluation
scfrank.bsky.social
Yes! It's the monolithic nature of the single value system that is the target of alignment that's so problematic. (But then we also have to agree to be ok with models that generate content that we as individuals are extremely un-aligned with, right?)
Reposted by Stella Frank
israsalazar.bsky.social
Today we are releasing Kaleidoscope 🎉

A comprehensive multimodal & multilingual benchmark for VLMs! It contains real questions from exams in different languages.

🌍 20,911 questions and 18 languages
📚 14 subjects (STEM → Humanities)
📸 55% multimodal questions
scfrank.bsky.social
The Panopticon is amazing! and thanks for this thread - my libby holds list just got a bit longer :-)
Reposted by Stella Frank
marcelbinz.bsky.social
We are looking for two PhD students at our institute in Munich.

Both postions are open-topic, so anything between cognitive science and machine learning is possible.

More information: hcai-munich.com/PhDHCAI.pdf

Feel free to share broadly!
hcai-munich.com
Reposted by Stella Frank
vlms4all.bsky.social
📢Excited to announce our upcoming workshop - Vision Language Models For All: Building Geo-Diverse and Culturally Aware Vision-Language Models (VLMs-4-All) @CVPR 2025!
🌐 sites.google.com/view/vlms4all
Reposted by Stella Frank
fgvcworkshop.bsky.social
BirdCLEF25: Audio-based species identification focused on birds, amphibians, mammals, and insects in Colombia.
👉 www.kaggle.com/competitions...
@cvprconference.bsky.social @kaggle.com
#FGVC #CVPR #CVPR2025 #LifeCLEF
[1/4]
Reposted by Stella Frank
aukeflorian.nl
Thanks to these insects, we can now study environmental microplastics retrospectively. 🔍 Even before Duprat began his now famous experiments with caddisfly larvae, insects in the wild were already experimenting with plastic... 🐛 14/x
Above: Casing of Ironoquia dubia (RMNH.INS.1544419) collected on May 18th 1971 in Loenen, The Netherlands. b) The label of the specimen.  Depicted on the right: detail of the artificial items.  Photographs: overview: Auke-Florian Hiemstra, details: Pasquale Ciliberti. Below: Caddisfly larvae in the studio of Hubert Duprat, carrying cases made from mostly gold. © Hubert Duprat, adagp, 2024, Courtesy the Artist and Art : Concept, Paris, Photo F. Delpech.
Reposted by Stella Frank
cameronwilson.bsky.social
The Wikimedia Foundation, which owns Wikipedia, says its bandwidth costs have gone up 50% since Jan 2024 — a rise they attribute to AI crawlers.

AI companies are killing the open web by stealing visitors from the sources of information and making them pay for the privilege