University of Washington
University of Washington
University of Washington
University of Washington
Galbot
University of Washington
University of Mannheim
Carnegie Mellon University
Microsoft Research
University of Washington
Carnegie Mellon University
*Equal contribution. Work done during internship at UW., †Corresponding author.
Vision-and-Language Navigation (VLN) systems often focus on either discrete (panoramic) or continuous (free-motion) paradigms alone, overlooking the complexities of human-populated, dynamic environments. We introduce a unified Human-Aware VLN (HA-VLN) benchmark that merges these paradigms under explicit social-awareness constraints. Our contributions include: (1) a standardized task definition that balances discrete-continuous navigation with personal-space requirements; (2) an enhanced human motion dataset (HAPS 2.0) and upgraded simulators capturing realistic multi-human interactions, outdoor contexts, and refined motion-language alignment; (3) extensive benchmarking on 16,844 human-centric instructions, revealing how multi-human dynamics and partial observability pose substantial challenges for leading VLN agents; (4) real-world robot tests validating sim-to-real transfer in crowded indoor spaces; and (5) a public leaderboard supporting transparent comparisons across discrete and continuous tasks. Empirical results show improved navigation success and fewer collisions when social context is integrated, underscoring the need for human-centric design. By releasing all datasets, simulators, agent code, and evaluation tools, we aim to advance safer, more capable, and socially responsible VLN research.
Single Humans with Movements (910 humans in total)
Visualization results of agent's trajectory
Examples of the robot navigating in different real environments.
@misc{dong2025havlnbenchmarkhumanawarenavigation,
title={HA-VLN: A Benchmark for Human-Aware Navigation in Discrete-Continuous Environments with Dynamic Multi-Human Interactions, Real-World Validation, and an Open Leaderboard},
author={Yifei Dong and Fengyi Wu and Qi He and Heng Li and Minghan Li and Zebang Cheng and Yuxuan Zhou and Jingdong Sun and Qi Dai and Zhi-Qi Cheng and Alexander G Hauptmann},
year={2025},
eprint={2503.14229},
archivePrefix={arXiv},
primaryClass={cs.AI},
url={https://arxiv.org/abs/2503.14229},
}