Learning pose estimation for UAV autonomous navigation and landing using visual-inertial sensor data
From Murray Wiki
Jump to navigationJump to search
Title | Learning pose estimation for UAV autonomous navigation and landing using visual-inertial sensor data |
---|---|
Authors | Francesca Baldini, Animashree Anandkumar and Richard M. Murray |
Source | 2020 American Control Conference (ACC) |
Abstract | In this work, we propose a robust network-in-the-loop control system for autonomous navigation and landing of an Unmanned-Aerial-Vehicle (UAV). To estimate the UAV's absolute pose, we develop a deep neural network (DNN) architecture for visual-inertial odometry, which provides a robust alternative to traditional methods. We first evaluate the accuracy of the estimation by comparing the prediction of our model to traditional visual-inertial approaches on the publicly available EuRoC MAV dataset. The results indicate a clear improvement in the accuracy of the pose estimation up to 25% over the baseline. Finally, we integrate the data-driven estimator in the closed-loop flight control system of Airsim, a simulator available as a plugin for Unreal Engine, and we provide simulation results for autonomous navigation and landing. |
Type | Conference paper |
URL | https://authors.library.caltech.edu/100568/1/1912.04527.pdf |
DOI | |
Tag | BAM20-acc |
ID | 2019i |
Funding | NSF VeHICaL |
Flags |