Sepideh Mamooler@ACL🇦🇹
@smamooler.bsky.social
39 followers 44 following 14 posts
PhD Candidate at @icepfl.bsky.social | Ex Research Intern@Google DeepMind 👩🏻‍💻 Working on multi-modal AI reasoning models in scientific domains https://smamooler.github.io/
Posts Media Videos Starter Packs
Pinned
smamooler.bsky.social
🚀 Introducing PICLe: a framework for in-context named-entity detection (NED) using pseudo-annotated demonstrations.
🎯 No human labeling needed—yet it outperforms few-shot learning with human annotations!
#AI #NLProc #LLMs #ICL #NER
smamooler.bsky.social
Also: not the usual #NLProc topic, but if you're working on genetic AI models, I’d love to connect! I'm exploring the intersection of behavioral genomics and multi-modal AI for behavior understanding.
smamooler.bsky.social
Excited to attend #ACL2025NLP in Vienna next week!
I’ll be presenting our #NAACL2025 paper, PICLe, at the first workshop on Large Language Models and Structure Modeling (XLLM) on Friday. Come by our poster if you’re into NER and ICL with pseudo-annotation.
arxiv.org/abs/2412.11923
PICLe: Pseudo-Annotations for In-Context Learning in Low-Resource Named Entity Detection
In-context learning (ICL) enables Large Language Models (LLMs) to perform tasks using few demonstrations, facilitating task adaptation when labeled examples are hard to obtain. However, ICL is sensiti...
arxiv.org
smamooler.bsky.social
Excited to attend #ACL2025NLP in Vienna next week!
I’ll be presenting our #NAACL2025 paper, PICLe, at the first workshop on Large Language Models and Structure Modeling (XLLM) on Friday. Come by our poster if you’re into NER and ICL with pseudo-annotation.
arxiv.org/abs/2412.11923
PICLe: Pseudo-Annotations for In-Context Learning in Low-Resource Named Entity Detection
In-context learning (ICL) enables Large Language Models (LLMs) to perform tasks using few demonstrations, facilitating task adaptation when labeled examples are hard to obtain. However, ICL is sensiti...
arxiv.org
Reposted by Sepideh Mamooler@ACL🇦🇹
bkhmsi.bsky.social
🚨 New Preprint!!

Thrilled to share with you our latest work: “Mixture of Cognitive Reasoners”, a modular transformer architecture inspired by the brain’s functional networks: language, logic, social reasoning, and world knowledge.

1/ 🧵👇
Reposted by Sepideh Mamooler@ACL🇦🇹
meltworkshop.bsky.social
🌍 ✨ Introducing MELT Workshop 2025: Multilingual, Multicultural, and Equitable Language Technologies
A workshop on building inclusive, culturally-aware LLMs!

🧠 Bridging the language divide in AI
📅 October 10, 2025 Co-located with @colmweb.org
🔗 melt-workshop.github.io

#MeltWorkshop2025
Reposted by Sepideh Mamooler@ACL🇦🇹
debjit-paul.bsky.social
Super excited to share that our paper "A Logical Fallacy-Informed Framework for Argument Generation" has received the Outstanding Paper Award 🎉🎉 at NAACL 2025!

Paper: aclanthology.org/2025.naacl-l...
Code: github.com/lucamouchel/...

#NAACL2025
smamooler.bsky.social
Couldn't attend @naaclmeeting.bsky.social in person as I didn't get a visa on time 🤷‍♀️ My colleague @mismayil.bsky.social will present PICLe on my behalf today, May 1st, at 3:15 pm in RUIDOSO. Feel free to reach out if you want to chat more!
smamooler.bsky.social
Check out VinaBench, our new #CVPR2025 paper. We introduce a benchmark for faithful and consistent visual narratives.

Paper: arxiv.org/abs/2503.20871
Project Page: silin159.github.io/Vina-Bench/
smamooler.bsky.social
Excited to share that we will present PICLe at @naaclmeeting.bsky.social main conference!
smamooler.bsky.social
🚀 Introducing PICLe: a framework for in-context named-entity detection (NED) using pseudo-annotated demonstrations.
🎯 No human labeling needed—yet it outperforms few-shot learning with human annotations!
#AI #NLProc #LLMs #ICL #NER
Reposted by Sepideh Mamooler@ACL🇦🇹
bkhmsi.bsky.social
🚨 New Preprint!!

LLMs trained on next-word prediction (NWP) show high alignment with brain recordings. But what drives this alignment—linguistic structure or world knowledge? And how does this alignment evolve during training? Our new paper explores these questions. 👇🧵
Reposted by Sepideh Mamooler@ACL🇦🇹
abosselut.bsky.social
Lots of great news out of the EPFL NLP lab these last few weeks. We'll be at @iclr-conf.bsky.social and @naaclmeeting.bsky.social in April / May to present some of our work in training dynamics, model representations, reasoning, and AI democratization. Come chat with us during the conference!
Reposted by Sepideh Mamooler@ACL🇦🇹
bkhmsi.bsky.social
🚨 New Paper!

Can neuroscience localizers uncover brain-like functional specializations in LLMs? 🧠🤖

Yes! We analyzed 18 LLMs and found units mirroring the brain's language, theory of mind, and multiple demand networks!

w/ @gretatuckute.bsky.social, @abosselut.bsky.social, @mschrimpf.bsky.social
🧵👇
smamooler.bsky.social
📊 We evaluate PICLe on 5 biomedical NED datasets and find:
✨ With zero human annotations, PICLe outperforms ICL in low-resource settings, where limited gold examples can be used as in-context demonstrations!
smamooler.bsky.social
⚙️ How does PICLe work?
1️⃣ LLMs annotate demonstrations in a zero-shot first pass.
2️⃣ Synthetic demos are clustered, and in-context sets are sampled.
3️⃣ Entity mentions are predicted using each set independently.
4️⃣ Self-verification selects the final predictions.
smamooler.bsky.social
💡 Building on our findings, we introduce PICLe: a framework for in-context learning powered by noisy, pseudo-annotated demonstrations. 🛠️ No human labels, no problem! 🚀
smamooler.bsky.social
📊 Key finding: A semantic mapping between demonstration context and label is essential for in-context task transfer. BUT even weak semantic mappings can provide enough signal for effective adaptation in NED!
smamooler.bsky.social
🔍 It’s unclear which demonstration attributes enable in-context learning in tasks that require structured, open-ended predictions (such as NED).
We use perturbation schemes that create demonstrations with varying correctness levels to analyze key demonstration attributes.
smamooler.bsky.social
🚀 Introducing PICLe: a framework for in-context named-entity detection (NED) using pseudo-annotated demonstrations.
🎯 No human labeling needed—yet it outperforms few-shot learning with human annotations!
#AI #NLProc #LLMs #ICL #NER