MAP: A Map-then-Act Paradigm for Long-Horizon Interactive Agent Reasoning
Quick Take
The MAP paradigm enhances interactive LLM agents by prioritizing environmental understanding before task execution.
Key Points
- MAP includes Global Exploration, Task-Specific Mapping, and Knowledge-Augmented Execution.
- Experiments show significant performance improvements across various benchmarks.
- MAP-2K dataset demonstrates that environment understanding surpasses imitation learning.
📖 Reader Mode
~2 min readAbstract:Current interactive LLM agents rely on goal-conditioned stepwise planning, where environmental understanding is acquired reactively during execution rather than established beforehand. This temporal inversion leads to Delayed Environmental Perception: agents must infer environmental constraints through trial-and-error, resulting in an Epistemic Bottleneck that traps them in inefficient failure cycles. Inspired by human affordance perception and cognitive map theory, we propose the Map-then-Act Paradigm (MAP), a plug-and-play framework that shifts environment understanding before execution. MAP consists of three stages: (1) Global Exploration, acquiring environment-general priors; (2) Task-Specific Mapping, constructing a structured cognitive map; and (3) Knowledge-Augmented Execution, solving tasks grounded on the map. Experiments show consistent gains across benchmarks and LLMs. On ARC-AGI-3, MAP enables frontier models to surpass near-zero baseline performance in 22 of 25 game environments. We further introduce MAP-2K, a dataset of map-then-act trajectories, and show that training on it outperforms expert execution traces, suggesting that understanding environments is more fundamental than imitation.
| Subjects: | Artificial Intelligence (cs.AI) |
| Cite as: | arXiv:2605.13037 [cs.AI] |
| (or arXiv:2605.13037v1 [cs.AI] for this version) | |
| https://doi.org/10.48550/arXiv.2605.13037 arXiv-issued DOI via DataCite (pending registration) |
Submission history
From: Yuxin Liu [view email]
[v1]
Wed, 13 May 2026 05:46:29 UTC (2,894 KB)
— Originally published at arxiv.org
More from arXiv cs.AI
See more →Invisible Orchestrators Suppress Protective Behavior and Dissociate Power-Holders: Safety Risks in Multi-Agent LLM Systems
Invisible orchestrators in multi-agent LLM systems pose significant safety risks and affect behavior dynamics.