SIGN: Safety-Aware Image-Goal Navigation for Autonomous Drones via Reinforcement Learning

Zichen Yan, Rui Huang, Lei He, Shao Guo, Lin Zhao
National University of Singapore

None


SIGN enables end-to-end ImageNav for autonomous drones via visual reinforcement learning with effective sim-to-real transfer. Guided by the collision predictor, an action correction mechanism is incorporated to ensure obstacle-avoiding navigation in indoor environments.

Abstract

Image-goal navigation (ImageNav) tasks a robot with autonomously exploring an unknown environment and reaching a location that visually matches a given target image. While prior works primarily study ImageNav for ground robots, enabling this capability for autonomous drones is substantially more challenging due to their need for high-frequency feedback control and global localization for stable flight. In this paper, we propose a novel sim-to-real framework that leverages reinforcement learning (RL) to achieve ImageNav for drones. To enhance visual representation ability, our approach trains the vision backbone with auxiliary tasks, including image perturbations and future transition prediction, which results in more effective policy training. The proposed algorithm enables end-to-end ImageNav with direct velocity control, eliminating the need for external localization. Furthermore, we integrate a depth-based safety module for real-time obstacle avoidance, allowing the drone to safely navigate in cluttered environments. Unlike most existing drone navigation methods that focus solely on reference tracking or obstacle avoidance, our framework supports comprehensive navigation behaviors, including autonomous exploration, obstacle avoidance, and image-goal seeking, without requiring explicit global mapping.

Simulation in Habitat

In-domain Evaluation: Gibson Benchmark

Cross-domain Evaluation: HM3D Benchmark

Cross-domain Evaluation: MP3D Benchmark

Real-world Experiments

Motion Capture Room

Office1

Office2