🔧 When environments change—say a new wall appears—ESWM adapts instantly. No retraining is needed. Just update the memory bank and the model replans.
This separation of memory and reasoning makes ESWM highly flexible.
🔧 When environments change—say a new wall appears—ESWM adapts instantly. No retraining is needed. Just update the memory bank and the model replans.
This separation of memory and reasoning makes ESWM highly flexible.
🧭 It gets even better!
ESWM can navigate between arbitrary points using only its memory bank—planning efficiently in latent space with near-optimal paths.
No access to global maps or coordinates required.
🧭 It gets even better!
ESWM can navigate between arbitrary points using only its memory bank—planning efficiently in latent space with near-optimal paths.
No access to global maps or coordinates required.
🚶 With no additional training, ESWM can explore novel environments efficiently by acting on uncertainty.
🚶 With no additional training, ESWM can explore novel environments efficiently by acting on uncertainty.
⚙️ How are these maps built?
We find that ESWM stitches together memories via overlapping states—merging local transitions into global structure.
Obstacles and boundaries serve as spatial anchors, guiding how memories are organized in latent space.
⚙️ How are these maps built?
We find that ESWM stitches together memories via overlapping states—merging local transitions into global structure.
Obstacles and boundaries serve as spatial anchors, guiding how memories are organized in latent space.
🏞️ How does ESWM solve the task?
Using ISOMAP, we visualize its latent representations—beautifully organized spatial layouts emerge from its internal states, even when the model sees only a small part or out-of-distribution environments.
🏞️ How does ESWM solve the task?
Using ISOMAP, we visualize its latent representations—beautifully organized spatial layouts emerge from its internal states, even when the model sees only a small part or out-of-distribution environments.
⚡️ Transformer-based ESWM models outperform LSTMs and Mamba, especially in settings where observations are compositional. Attention allows the model to flexibly bind relevant memories and generalize across structures.
⚡️ Transformer-based ESWM models outperform LSTMs and Mamba, especially in settings where observations are compositional. Attention allows the model to flexibly bind relevant memories and generalize across structures.
To train ESWM, we use meta-learning across diverse environments. At test time, the model gets a minimal set of disjoint episodic memories (single transitions) and must predict a missing element in a new transition—without ever seeing the full map.
To train ESWM, we use meta-learning across diverse environments. At test time, the model gets a minimal set of disjoint episodic memories (single transitions) and must predict a missing element in a new transition—without ever seeing the full map.
🧠 Inspired by the MTL’s architecture and function, we built ESWM: a neural network that infers the structure of its environment from isolated, one-step transitions—just like the brain integrates episodes into a cognitive map.
🧠 Inspired by the MTL’s architecture and function, we built ESWM: a neural network that infers the structure of its environment from isolated, one-step transitions—just like the brain integrates episodes into a cognitive map.
We introduce the Episodic Spatial World Model (ESWM)—a model that constructs flexible internal world models from sparse, disjoint memories.
🧵👇 [1/12]
We introduce the Episodic Spatial World Model (ESWM)—a model that constructs flexible internal world models from sparse, disjoint memories.
🧵👇 [1/12]