Hierarchical Instruction-aware Embodied Visual Tracking
Kui Wu, Hao Chen, Churan Wang, Fakhri Karray, Zhoujun Li, Yizhou Wang, Fangwei Zhong
Unverified — Be the first to reproduce this paper.
ReproduceAbstract
User-Centric Embodied Visual Tracking (UC-EVT) presents a novel challenge for reinforcement learning-based models due to the substantial gap between high-level user instructions and low-level agent actions. While recent advancements in language models (e.g., LLMs, VLMs, VLAs) have improved instruction comprehension, these models face critical limitations in either inference speed (LLMs, VLMs) or generalizability (VLAs) for UC-EVT tasks. To address these challenges, we propose Hierarchical Instruction-aware Embodied Visual Tracking (HIEVT) agent, which bridges instruction comprehension and action generation using spatial goals as intermediaries. HIEVT first introduces LLM-based Semantic-Spatial Goal Aligner to translate diverse human instructions into spatial goals that directly annotate the desired spatial position. Then the RL-based Adaptive Goal-Aligned Policy, a general offline policy, enables the tracker to position the target as specified by the spatial goal. To benchmark UC-EVT tasks, we collect over ten million trajectories for training and evaluate across one seen environment and nine unseen challenging environments. Extensive experiments and real-world deployments demonstrate the robustness and generalizability of HIEVT across diverse environments, varying target dynamics, and complex instruction combinations. The complete project is available at https://sites.google.com/view/hievt.