Tal Linzen
@tallinzen.bsky.social
2.9K followers 79 following 19 posts
NYU professor, Google research scientist. Good at LaTeX.
Posts Media Videos Starter Packs
Reposted by Tal Linzen
luckytran.com
DO NOT GIVE UP!

Our advocacy is working.

A key Senate committee has indicated that it will reject Trump’s proposed cuts to science agencies including NASA and the NSF.

Keep speaking up and calling your electeds 🗣️🗣️🗣️
Nature: US senators poised to reject Trump’s proposed massive science cuts

Committee gives first hint that policymakers might preserve, rather than slash, funding for US National Science Foundation and other agencies.
tallinzen.bsky.social
Maybe five years with a no-cost extension!
tallinzen.bsky.social
If we have a lot of shared followers, perhaps you could comment on the pinned tweet on my account and provide context?! Thank you!
tallinzen.bsky.social
My Twitter account has been hacked :( Please don't click on any links "I" posted on that account recently!
tallinzen.bsky.social
I'll be accepting applications for a while, and will also consider people with a late start date. Feel free to email if you have questions. No need for a formal cover letter.
tallinzen.bsky.social
The goal is to model some cool behavioral and neural data from humans (some to be collected) but we expect to do a lot of fundamental modeling and interpretability work. You don't need to have existing experience in cognitive science but you should be interested in learning more about it.
tallinzen.bsky.social
I'm hiring at least one post-doc! We're interested in creating language models that process language more like humans than mainstream LLMs do, through architectural modifications and interpretability-style steering. Express interest here: docs.google.com/forms/d/e/1F...
NYU LLM + cognitive science post-doc interest form
Tal Linzen's group at NYU is hiring a post-doc! We're interested in creating language models that process language more like humans than mainstream LLMs do, through architectural modifications and int...
docs.google.com
tallinzen.bsky.social
דווקא בביקורת הדרכונים בצד הישראלי של המעבר לא ביקשו הוכחה שיש לי גם דרכון זר, נתנו לי לצאת מהארץ בלי בעייה. מסתבר שיש סיכון בטחוני ביציאה מהארץ של אזרחים ישראלים רק בדרך האוויר, לא בדרך היבשה
Reposted by Tal Linzen
jacksonpetty.org
How well can LLMs understand tasks with complex sets of instructions? We investigate through the lens of RELIC: REcognizing (formal) Languages In-Context, finding a significant overhang between what LLMs are able to do theoretically and how well they put this into practice.
Reposted by Tal Linzen
arianna-bis.bsky.social
Following the success story of BabyBERTa, I & many other NLPers have turned to language acquisition for inspiration. In this new paper we show that using Child-Directed Language as training data is unfortunately *not* beneficial for syntax learning, at least not in the traditional LM training regime
tallinzen.bsky.social
Depends on what you mean by US academics, I guess. A lot of people are here for a temporary position, don't have strong ties to the country, and were mentally prepared to move elsewhere anyway. Those people are much more likely to leave than before.
tallinzen.bsky.social
I'll have a bit of time to chat with folks in Berlin and/or Copenhagen about AI, LLMs, cognitive science, how good your bike infrastructure is, etc, let me know!
tallinzen.bsky.social
And this one on language models with cognitively plausible memory in Potsdam on Tuesday (as part of this in-person-only sentence processing workshop vasishth.github.io/sentproc-wor...):
tallinzen.bsky.social
Cross-posting the abstracts for two talks I'm giving next week! This one on formal languages for LLM pretraining and evaluation, at Apple ML Research in Copenhagen on Wednesday
tallinzen.bsky.social
Updated version of our position piece on how language models can help us understand how people learn and process language, on why it's crucial to train models on cognitive plausible datasets, and on the BabyLM project that addresses this issue.
wegotlieb.bsky.social
📣Paper Update 📣It’s bigger! It’s better! Even if the language models aren’t. 🤖New version of “Bigger is not always Better: The importance of human-scale language modeling for psycholinguistics” osf.io/preprints/ps...
OSF
osf.io
tallinzen.bsky.social
out of date, should be $300 billion now!
tallinzen.bsky.social
thanks! I'll start with the frens and nice people and work my way up from there!
Reposted by Tal Linzen
dariopaape.bsky.social
At #HSP2025, I'll present work with @tallinzen.bsky.social and @shravanvasishth.bsky.social on modeling garden-pathing in a huge benchmark dataset: hsp2025.github.io/abstracts/29.... Statistically decomposing the effect into subprocesses greatly improves predictive fit over just comparing means!
hsp2025.github.io
tallinzen.bsky.social
Going to give this website another shot! What are good lists of linguistics, psycholinguistics, NLP and AI accounts?
tallinzen.bsky.social
Thanks Ted for mentioning me in the same tweet as Chris! This website really is better than the other one!
tedunderwood.com
Another good sign for this platform is that people like @tallinzen.bsky.social and @chrmanning.bsky.social are dipping their toes in it (even if no pfps yet). Give them a follow if you know who they are.
tallinzen.bsky.social
Very little happening on here but silence is certainly better than all of the boardroom drama takes on the other website. Four different people I follow just came up with the same unfunny joke about the most recent development in the drama, apparently independently?
Reposted by Tal Linzen
jacksonpetty.org
Do deep transformer LMs generalize better? In a new preprint we (Sjoerd van Steenkiste, Ishita Dasgupta, Fei Sha, Dan Garrette, & @tallinzen.bsky.social) control for parameter count to show how depth helps models on compositional generalization tasks, but diminishingly so 🧵

jacksonpetty.org/depth
The Impact of Depth and Width on Transformer Language Model Generalization
To process novel sentences, language models (LMs) must generalize compositionally -- combine familiar elements in new ways. What aspects of a model's structure promote compositional generalization? Fo...
jacksonpetty.org