Jiaang Li
@jiaangli.bsky.social
1.1K followers 85 following 17 posts
PhD student at University of Copenhagen @belongielab.org | #nlp #computervision | ELLIS student @ellis.eu 🌐 https://jiaangli.github.io/
Posts Media Videos Starter Packs
jiaangli.bsky.social
Feel free to reach out and chat with Xinyi on July 18th in Vancouver at the #ICML
xinyichen2024.bsky.social
Excited to present at the #ICML2025 World Models Workshop!
📅 July 18, 15:45–17:00
🧠 What if Othello-Playing Language Models Could See?
We show that visual grounding improves prediction & internal structure.♟️
Reposted by Jiaang Li
serge.belongie.com
Would you present your next NeurIPS paper in Europe instead of traveling to San Diego (US) if this was an option? Søren Hauberg (DTU) and I would love to hear the answer through this poll: (1/6)
NeurIPS participation in Europe
We seek to understand if there is interest in being able to attend NeurIPS in Europe, i.e. without travelling to San Diego, US. In the following, assume that it is possible to present accepted papers ...
docs.google.com
Reposted by Jiaang Li
sloeschcke.bsky.social
Check out our new preprint 𝐓𝐞𝐧𝐬𝐨𝐫𝐆𝐑𝐚𝐃.
We use a robust decomposition of the gradient tensors into low-rank + sparse parts to reduce optimizer memory for Neural Operators by up to 𝟕𝟓%, while matching the performance of Adam, even on turbulent Navier–Stokes (Re 10e5).
Reposted by Jiaang Li
aicentre.dk
PhD student, Jiaang Li and his collaborators, with insights into cultural understanding of vision-language models 👇
jiaangli.bsky.social
🚀New Preprint🚀
Can Multimodal Retrieval Enhance Cultural Awareness in Vision-Language Models?

Excited to introduce RAVENEA, a new benchmark aimed at evaluating cultural understanding in VLMs through RAG.
arxiv.org/abs/2505.14462

More details:👇
Reposted by Jiaang Li
srishtiy.bsky.social
I am excited to announce our latest work 🎉 "Cultural Evaluations of Vision-Language Models Have a Lot to Learn from Cultural Theory". We review recent works on culture in VLMs and argue for deeper grounding in cultural theory to enable more inclusive evaluations.

Paper 🔗: arxiv.org/pdf/2505.22793
Paper title "Cultural Evaluations of Vision-Language Models
Have a Lot to Learn from Cultural Theory"
jiaangli.bsky.social
Great collaboration with @yfyuan01.bsky.social @wenyan62.bsky.social @aliannejadi.bsky.social @danielhers.bsky.social , Anders Søgaard, Ivan Vulić, Wenxuan Zhang, Paul Liang, Yang Deng, @serge.belongie.com
jiaangli.bsky.social
📊Our experiments demonstrate that even lightweight VLMs, when augmented with culturally relevant retrievals, outperform their non-augmented counterparts and even surpass the next larger model tier, achieving at least a 3.2% improvement in cVQA and 6.2% in cIC.
jiaangli.bsky.social
🛠Culture-Aware Contrastive Learning

We propose Culture-aware Contrastive (CAC) Learning, a supervised learning framework compatible with both CLIP and SigLIP architectures. Fine-tuning with CAC can help models better capture culturally significant content.
jiaangli.bsky.social
📚 Dataset Construction
RAVENEA integrates 1,800+ images, 2,000+ culture-related questions, 500+ human captions, and 10,000+ human-ranked Wikipedia documents to support two key tasks:

🎯Culture-focused Visual Question Answering (cVQA)
📝Culture-informed Image Captioning (cIC)
jiaangli.bsky.social
🚀New Preprint🚀
Can Multimodal Retrieval Enhance Cultural Awareness in Vision-Language Models?

Excited to introduce RAVENEA, a new benchmark aimed at evaluating cultural understanding in VLMs through RAG.
arxiv.org/abs/2505.14462

More details:👇
jiaangli.bsky.social
Super cool! Incidentally, in our previous project, we also found that linear alignment between embedding spaces from two modalities is viable — and the alignment improves as LLMs scale.
bsky.app/profile/jiaa...
jiaangli.bsky.social
🤔Do Vision and Language Models Share Concepts? 🚀
We present an empirical evaluation and find that language models partially converge towards representations isomorphic to those of vision models. #EMNLP

📃 direct.mit.edu/tacl/article...
Reposted by Jiaang Li
yfyuan01.bsky.social
I won’t be attending #ICLR in person this year😢. But feel free to check our paper ‘Revisiting the Othello World Model Hypothesis’ with Anders Søgaard, accepted at ICLR world models workshop!
Paper link arxiv.org/abs/2503.04421
Revisiting the Othello World Model Hypothesis
Li et al. (2023) used the Othello board game as a test case for the ability of GPT-2 to induce world models, and were followed up by Nanda et al. (2023b). We briefly discuss the original experiments, ...
arxiv.org
Reposted by Jiaang Li
zhaochongan.bsky.social
Thrilled to announce "Multimodality Helps Few-shot 3D Point Cloud Semantic Segmentation" is accepted as a Spotlight (5%) at #ICLR2025!

Our model MM-FSS leverages 3D, 2D, & text modalities for robust few-shot 3D segmentation—all without extra labeling cost. 🤩

arxiv.org/pdf/2410.22489

More details👇
Reposted by Jiaang Li
chengzu-li.bsky.social
Forget just thinking in words.

🔔Our New Preprint:
🚀 New Era of Multimodal Reasoning🚨
🔍 Imagine While Reasoning in Space with MVoT

Multimodal Visualization-of-Thought (MVoT) revolutionizes reasoning by generating visual "thoughts" that transform how AI thinks, reasons, and explains itself.
Reposted by Jiaang Li
nicolang.bsky.social
FGVC12 Workshop is coming to #CVPR 2025 in Nashville!

Are you working on fine-grained visual problems?
This year we have two peer-reviewed paper tracks:
i) 8-page CVPR Workshop proceedings
ii) 4-page non-archival extended abstracts
CALL FOR PAPERS: sites.google.com/view/fgvc12/...
Reposted by Jiaang Li
serge.belongie.com
Here’s a short film produced by the Danish Royal Academy of Sciences, showcasing the WineSensed 🍷 project of Þóranna Bender et al. thoranna.github.io/learning_to_...
VidenSkaber | Min AI forstår mig ikke - professor Serge Belongie
YouTube video by Videnskabernes Selskab
youtu.be
Reposted by Jiaang Li
belongielab.org
From San Diego to New York to Copenhagen, wishing you Happy Holidays!🎄
Reposted by Jiaang Li
belongielab.org
With @neuripsconf.bsky.social right around the corner, we’re excited to be presenting our work soon! Here’s an overview

(1/5)
Reposted by Jiaang Li
belongielab.org
Here’s a starter pack with members of our lab that have joined Bluesky
Belongie Lab
Join the conversation
go.bsky.app
jiaangli.bsky.social
🚀Take away:

1. Representation spaces of LMs and VMs grow more partially similar with model size.
2. Lower frequency, polysemy, dispersion can be easier to align.
3. Shared concepts between LMs and VMs might extend beyond nouns.

🧵(7/8)
#NLP #NLProc
jiaangli.bsky.social
🌱We then discuss the implications of our finding:
- the LM understanding debate
- the study of emergent properties
- philosophy

🧵(6/8)