Ethan Gotlieb Wilcox
@wegotlieb.bsky.social
880 followers 180 following 20 posts
Assistant Professor of Computational Linguistics @ Georgetown; formerly postdoc @ ETH Zurich; PhD @ Harvard Linguistics, affiliated with MIT Brain & Cog Sci. Language, Computers, Cognition.
Posts Media Videos Starter Packs
wegotlieb.bsky.social
🌟🌟This paper will appear at ACL 2025 (@aclmeeting.bsky.social)! New updated version is on arXiv: arxiv.org/pdf/2505.07659 🌟🌟
Reposted by Ethan Gotlieb Wilcox
sashaboguraev.bsky.social
A key hypothesis in the history of linguistics is that different constructions share underlying structure. We take advantage of recent advances in mechanistic interpretability to test this hypothesis in Language Models.

New work with @kmahowald.bsky.social and @cgpotts.bsky.social!

🧵👇!
wegotlieb.bsky.social
✅In line with our prediction, we find that mutual information is higher in tonal languages than in non-tonal languages. BUT, the way one represents context is important. When full sentential context is taken into account (mBERT and mGPT), the distinction collapses.
wegotlieb.bsky.social
🌏🌍We test this prediction by estimating mutual information in an audio dataset of 10 different languages across 6 language families. 🌏🌍
wegotlieb.bsky.social
We propose a way to do so using …📡information theory.📡 In tonal languages, pitch reduces uncertainty about lexical identity, therefore, the mutual information between pitch and words should be higher.
wegotlieb.bsky.social
🌐But there are intermediate languages, which have lexically contrastive tone, but only sporadically, making some linguists doubt the tonal/non-tonal dichotomy. So, how can we measure how “tonal” a language is? 🧐🧐
wegotlieb.bsky.social
🌏 Different languages use pitch in different ways. 🌏 “Tonal” languages, like Cantonese, use it to make lexical distinctions. 📖 While others, like English, use it for other functions, like marking whether or not a sentence is a question. ❓
wegotlieb.bsky.social
I’ll also use this as a way to plug human-scale language modeling in the wild: This year’s BabyLM eval pipeline was just released last week at github.com/babylm/evalu.... For more info on BabyLM head to babylm.github.io
GitHub - babylm/evaluation-pipeline-2025
Contribute to babylm/evaluation-pipeline-2025 development by creating an account on GitHub.
github.com
wegotlieb.bsky.social
Couldn’t be happier to have co-authored this will a stellar team, including: Michael Hu, @amuuueller.bsky.social, @alexwarstadt.bsky.social, @lchoshen.bsky.social, Chengxu Zhuang, @adinawilliams.bsky.social, Ryan Cotterell, @tallinzen.bsky.social
wegotlieb.bsky.social
This version includes 😱New analyses 😱new arguments 😱 and a whole new “Looking Forward” section! If you’re interested in what a team of (psycho) computational linguists thinks the future will hold, check out our brand new Section 8!
wegotlieb.bsky.social
📣Paper Update 📣It’s bigger! It’s better! Even if the language models aren’t. 🤖New version of “Bigger is not always Better: The importance of human-scale language modeling for psycholinguistics” osf.io/preprints/ps...
OSF
osf.io
Reposted by Ethan Gotlieb Wilcox
cuiding.bsky.social
Excited to share our preprint "Using MoTR to probe agreement errors in Russian"! w/ Metehan Oğuz, @wegotlieb.bsky.social, Zuzanna Fuchs Link: osf.io/preprints/ps...
1- We provide moderate evidence that processing of agreement errors is modulated by agreement type (internal vs external agr.)
OSF
osf.io
Reposted by Ethan Gotlieb Wilcox
johnbasil.bsky.social
Me and @wegotlieb.bsky.social were recently invited to write a wide-ranging reflection on the current state of linguistic theory and methodology.
A draft is up here. For anyone interested in thinking big about linguistics, we'd be happy to hear your thoughts!
arxiv.org/abs/2502.18313
#linguistics
Looking forward: Linguistic theory and methods
This chapter examines current developments in linguistic theory and methods, focusing on the increasing integration of computational, cognitive, and evolutionary perspectives. We highlight four major ...
arxiv.org
wegotlieb.bsky.social
⚖️📣This paper was a big departure from my typical cognitive science fare, and so much fun to write! 📣⚖️ Thank you to @bwal.bsky.social and especially to @kevintobia.bsky.social for their legal expertise on this project!
wegotlieb.bsky.social
On the positive side, we suggest that LLMs can serve a role as “dialectic” partners 🗣️❔🗣️ helping judges and clerks strengthen their arguments, as long as judicial sovereignty is maintained 👩‍⚖️👑👩‍⚖️
wegotlieb.bsky.social
⚖️ We also show, through demonstration, that it’s very easy to engineer prompts that steer models toward one’s desired interpretation of a word or phrase. 📖Prompting is the new “dictionary shopping” 😬 📖 😬
wegotlieb.bsky.social
🏛️We identify five “myths” about LLMs which, when dispelled, reveal their limitations as legal tools for textual interpretation. To take one example, during instruction tuning, LLMs are trained on highly structured, non-natural inputs.
wegotlieb.bsky.social
We argue no! 🙅‍♂️ While LLMs appear to possess excellent language capabilities, they should not be used as references for “ordinary language use,” at least in the legal setting. ⚖️ The reasons are manifold.
wegotlieb.bsky.social
🏛️Last year a U.S. judge queried Chat GPT to help with their interpretation of “ordinary meaning,” in the same way one might use a dictionary to look up the ordinary definition of a word 📖 … But is it the same?