Martin Engelcke
martinengelcke.bsky.social
Martin Engelcke
@martinengelcke.bsky.social
Senior Research Scientist at Google DeepMind. Views my own.
Pinned
The technical report with more details about our work on SIMA 2 is now also available on arXiv!

Take a look at: arxiv.org/abs/2512.04797
SIMA 2: A Generalist Embodied Agent for Virtual Worlds
We introduce SIMA 2, a generalist embodied agent that understands and acts in a wide variety of 3D virtual worlds. Built upon a Gemini foundation model, SIMA 2 represents a significant step toward act...
arxiv.org
Reposted by Martin Engelcke
Hello! 👋

Are you interested in AI for board games using language models? Want to do some hobby tinkering with fine-tuning or RL?

We've released an easy-to-follow example colab that fine-tunes Gemma models via Kauldron to mimic an MCTS player.

Details here: github.com/google-deepm...

♟️🎲♦️♠️♥️♣️✨🎉
2025 Wrap-up: Fine-tuning Gemma with Kauldron Example ✦︎ · Issue #1414 · google-deepmind/open_spiel
Hello everyone! We've been hard at work this year working on OpenSpiel 2.0, which will be better than ever. Major developments have been underway to make working with language models easier. I'm lo...
github.com
December 19, 2025 at 6:35 PM
Reposted by Martin Engelcke
Why isn’t modern AI built around principles from cognitive science or neuroscience? Starting a substack (infinitefaculty.substack.com/p/why-isnt-m...) by writing down my thoughts on that question: as part of a first series of posts giving my current thoughts on the relation between these fields. 1/3
Why isn’t modern AI built around principles from cognitive science?
First post in a series on cognitive science and AI
infinitefaculty.substack.com
December 16, 2025 at 3:40 PM
Reposted by Martin Engelcke
What are you favourite Imitation Learning and Inverse Reinforcement Learning papers?
What are the essential papers?

It doesn't matter much whether they are new or old, but I prefer a conceptually elegant and mathematically solid work.
December 16, 2025 at 12:33 AM
The technical report with more details about our work on SIMA 2 is now also available on arXiv!

Take a look at: arxiv.org/abs/2512.04797
SIMA 2: A Generalist Embodied Agent for Virtual Worlds
We introduce SIMA 2, a generalist embodied agent that understands and acts in a wide variety of 3D virtual worlds. Built upon a Gemini foundation model, SIMA 2 represents a significant step toward act...
arxiv.org
December 10, 2025 at 3:13 PM
Reposted by Martin Engelcke
Nice tweet thread from @dannypsawyer on work exploring how well frontier models like GPT, Claude, and Gemini explore in interactive, multi-turn settings - to be presented at NeurIPS workshops this December!
Happy to announce that our work has been accepted to workshops on Multi-turn Interactions and Embodied World Models at #NeurIPS2025! Frontier foundation models are incredible, but how well can they explore in interactive environments?
Paper👇
arxiv.org/abs/2412.06438
🧵1/13
October 10, 2025 at 5:18 PM
Reposted by Martin Engelcke
Happy to announce that our work has been accepted to workshops on Multi-turn Interactions and Embodied World Models at #NeurIPS2025! Frontier foundation models are incredible, but how well can they explore in interactive environments?
Paper👇
arxiv.org/abs/2412.06438
🧵1/13
October 10, 2025 at 5:11 PM
Our work on "Latent learning: episodic memory complements parametric learning by enabling flexible reuse of experiences" led by @lampinen.bsky.social and with Effie Li, @arslanchaudhry.bsky.social, and James McClelland is now available on arXiv!

Link: arxiv.org/abs/2509.16189

Thread: 1/
Why does AI sometimes fail to generalize, and what might help? In a new paper (arxiv.org/abs/2509.16189), we highlight the latent learning gap — which unifies findings from language modeling to agent navigation — and suggest that episodic memory complements parametric learning to bridge it. Thread:
Latent learning: episodic memory complements parametric learning by enabling flexible reuse of experiences
When do machine learning systems fail to generalize, and what mechanisms could improve their generalization? Here, we draw inspiration from cognitive science to argue that one weakness of machine lear...
arxiv.org
September 29, 2025 at 11:02 AM
Hello, World!
July 27, 2025 at 10:55 AM