TY - JOUR
T1 - Fusing Vision and Inertial Sensors for Robust Runway Detection and Tracking
AU - Abu Jbara, Khaled F.
AU - Sundaramoorthi, Ganesh
AU - Claudel, Christian
N1 - KAUST Repository Item: Exported on 2020-10-01
Acknowledgements: We gratefully acknowledge the Aerospace Division of King Abdulaziz City for Science and Technology (KACST) for supporting us with experimental flight videos generated by the SAKER 4 UAV.
PY - 2018/6/30
Y1 - 2018/6/30
N2 - This work presents a novel real-time algorithm for runway detection and tracking applied to unmanned aerial vehicles (UAVs). The algorithm relies on a combination of segmentation-based region competition and minimization of a particular energy function to detect and identify the runway edges from streaming video data. The resulting video-based runway position estimates can be updated using a Kalman filter (KF) that integrates additional kinematic estimates such as position and attitude angles, derived from video, inertial measurement unit data, or positioning data. This allows a more robust tracking of the runway under turbulence. The performance of the proposed lane detection and tracking scheme is illustrated on various experimental UAV flights conducted by the Saudi Aerospace Research Center (KACST), by the University of Texas, Austin, and on simulated landing videos obtained from a flight simulator. Results show an accurate tracking of the runway edges during the landing phase, under various lighting conditions, even in the presence of roads, taxiways, and other obstacles. This suggests that the positional estimates derived from the video data can significantly improve the guidance of the UAV during takeoff and landing phases.
AB - This work presents a novel real-time algorithm for runway detection and tracking applied to unmanned aerial vehicles (UAVs). The algorithm relies on a combination of segmentation-based region competition and minimization of a particular energy function to detect and identify the runway edges from streaming video data. The resulting video-based runway position estimates can be updated using a Kalman filter (KF) that integrates additional kinematic estimates such as position and attitude angles, derived from video, inertial measurement unit data, or positioning data. This allows a more robust tracking of the runway under turbulence. The performance of the proposed lane detection and tracking scheme is illustrated on various experimental UAV flights conducted by the Saudi Aerospace Research Center (KACST), by the University of Texas, Austin, and on simulated landing videos obtained from a flight simulator. Results show an accurate tracking of the runway edges during the landing phase, under various lighting conditions, even in the presence of roads, taxiways, and other obstacles. This suggests that the positional estimates derived from the video data can significantly improve the guidance of the UAV during takeoff and landing phases.
UR - http://hdl.handle.net/10754/628829
UR - https://arc.aiaa.org/doi/10.2514/1.G002898
UR - http://www.scopus.com/inward/record.url?scp=85052883015&partnerID=8YFLogxK
U2 - 10.2514/1.G002898
DO - 10.2514/1.G002898
M3 - Article
SN - 0731-5090
VL - 41
SP - 1929
EP - 1946
JO - Journal of Guidance, Control, and Dynamics
JF - Journal of Guidance, Control, and Dynamics
IS - 9
ER -