From reactive to cognitive: brain-inspired spatial intelligence for embodied agents
Abstract
BSC-Nav constructs allocentric cognitive maps from egocentric trajectories and contextual cues, enabling embodied agents to perform diverse navigation tasks with zero-shot generalization and versatile behaviors.
Spatial cognition enables adaptive goal-directed behavior by constructing internal models of space. Robust biological systems consolidate spatial knowledge into three interconnected forms: landmarks for salient cues, route knowledge for movement trajectories, and survey knowledge for map-like representations. While recent advances in multi-modal large language models (MLLMs) have enabled visual-language reasoning in embodied agents, these efforts lack structured spatial memory and instead operate reactively, limiting their generalization and adaptability in complex real-world environments. Here we present Brain-inspired Spatial Cognition for Navigation (BSC-Nav), a unified framework for constructing and leveraging structured spatial memory in embodied agents. BSC-Nav builds allocentric cognitive maps from egocentric trajectories and contextual cues, and dynamically retrieves spatial knowledge aligned with semantic goals. Integrated with powerful MLLMs, BSC-Nav achieves state-of-the-art efficacy and efficiency across diverse navigation tasks, demonstrates strong zero-shot generalization, and supports versatile embodied behaviors in the real physical world, offering a scalable and biologically grounded path toward general-purpose spatial intelligence.
Community
From reactive to cognitive: brain-inspired spatial intelligence for embodied agents
This work introduces BSC-Nav, a brain-inspired spatial intelligence framework. By explicitly constructing and leveraging a structured spatial memory composed of landmarks, route knowledge, and survey knowledge, BSC-Nav shifts agents from reactive behavior to cognitive navigation, demonstrating strong zero-shot generalization in complex real-world tasks, including universal navigation, mobile manipulation, and spatial reasoning through active exploration.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- $NavA^3$: Understanding Any Instruction, Navigating Anywhere, Finding Anything (2025)
- Imaginative World Modeling with Scene Graphs for Embodied Agent Navigation (2025)
- GoViG: Goal-Conditioned Visual Navigation Instruction Generation (2025)
- EmbRACE-3K: Embodied Reasoning and Action in Complex Environments (2025)
- OpenNav: Open-World Navigation with Multimodal Large Language Models (2025)
- Spatial Policy: Guiding Visuomotor Robotic Manipulation with Spatial-Aware Modeling and Reasoning (2025)
- CogDDN: A Cognitive Demand-Driven Navigation with Decision Optimization and Dual-Process Thinking (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper