TY - GEN
T1 - UAV Path Planning using Global and Local Map Information with Deep Reinforcement Learning
AU - Theile, Mirco
AU - Bayerlein, Harald
AU - Nai, Richard
AU - Gesbert, David
AU - Caccamo, Marco
N1 - Publisher Copyright:
© 2021 IEEE.
PY - 2021
Y1 - 2021
N2 - Path planning methods for autonomous unmanned aerial vehicles (UAVs) are typically designed for one specific type of mission. This work presents a method for autonomous UAV path planning based on deep reinforcement learning (DRL) that can be applied to a wide range of mission scenarios. Specifically, we compare coverage path planning (CPP), where the UAV's goal is to survey an area of interest to data harvesting (DH), where the UAV collects data from distributed Internet of Things (IoT) sensor devices. By exploiting structured map information of the environment, we train double deep Q-networks (DDQNs) with identical architectures on both distinctly different mission scenarios to make movement decisions that balance the respective mission goal with navigation constraints. By introducing a novel approach exploiting a compressed global map of the environment combined with a cropped but uncompressed local map showing the vicinity of the UAV agent, we demonstrate that the proposed method can efficiently scale to large environments. We also extend previous results for generalizing control policies that require no retraining when scenario parameters change and offer a detailed analysis of crucial map processing parameters' effects on path planning performance.
AB - Path planning methods for autonomous unmanned aerial vehicles (UAVs) are typically designed for one specific type of mission. This work presents a method for autonomous UAV path planning based on deep reinforcement learning (DRL) that can be applied to a wide range of mission scenarios. Specifically, we compare coverage path planning (CPP), where the UAV's goal is to survey an area of interest to data harvesting (DH), where the UAV collects data from distributed Internet of Things (IoT) sensor devices. By exploiting structured map information of the environment, we train double deep Q-networks (DDQNs) with identical architectures on both distinctly different mission scenarios to make movement decisions that balance the respective mission goal with navigation constraints. By introducing a novel approach exploiting a compressed global map of the environment combined with a cropped but uncompressed local map showing the vicinity of the UAV agent, we demonstrate that the proposed method can efficiently scale to large environments. We also extend previous results for generalizing control policies that require no retraining when scenario parameters change and offer a detailed analysis of crucial map processing parameters' effects on path planning performance.
UR - http://www.scopus.com/inward/record.url?scp=85124685476&partnerID=8YFLogxK
U2 - 10.1109/ICAR53236.2021.9659413
DO - 10.1109/ICAR53236.2021.9659413
M3 - Conference contribution
AN - SCOPUS:85124685476
T3 - 2021 20th International Conference on Advanced Robotics, ICAR 2021
SP - 539
EP - 546
BT - 2021 20th International Conference on Advanced Robotics, ICAR 2021
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 20th International Conference on Advanced Robotics, ICAR 2021
Y2 - 6 December 2021 through 10 December 2021
ER -