TY - GEN
T1 - Fast and accurate trajectory tracking for unmanned aerial vehicles based on deep reinforcement learning
AU - Li, Yilan
AU - Li, Hongjia
AU - Li, Zhe
AU - Fang, Haowen
AU - Sanyal, Amit K.
AU - Wang, Yanzhi
AU - Qiu, Qinru
N1 - Publisher Copyright:
© 2019 IEEE.
PY - 2019/8
Y1 - 2019/8
N2 - Continuous trajectory control of fixed-wing unmanned aerial vehicles (UAVs) is complicated when considering hidden dynamics. Due to UAV multi degrees of freedom, tracking methodologies based on conventional control theory, such as Proportional-Integral-Derivative (PID) has limitations in response time and adjustment robustness, while a model based approach that calculates the force and torques based on UAV's current status is complicated and rigid. We present an actor-critic reinforcement learning framework that controls UAV trajectory through a set of desired waypoints. A deep neural network is constructed to learn the optimal tracking policy and reinforcement learning is developed to optimize the resulting tracking scheme. The experimental results show that our proposed approach can achieve 58.14% less position error, 21.77% less system power consumption and 9.23% faster attainment than the baseline. The actor network consists of only linear operations, hence Field Programmable Gate Arrays (FPGA) based hardware acceleration can easily be designed for energy efficient real-time control.
AB - Continuous trajectory control of fixed-wing unmanned aerial vehicles (UAVs) is complicated when considering hidden dynamics. Due to UAV multi degrees of freedom, tracking methodologies based on conventional control theory, such as Proportional-Integral-Derivative (PID) has limitations in response time and adjustment robustness, while a model based approach that calculates the force and torques based on UAV's current status is complicated and rigid. We present an actor-critic reinforcement learning framework that controls UAV trajectory through a set of desired waypoints. A deep neural network is constructed to learn the optimal tracking policy and reinforcement learning is developed to optimize the resulting tracking scheme. The experimental results show that our proposed approach can achieve 58.14% less position error, 21.77% less system power consumption and 9.23% faster attainment than the baseline. The actor network consists of only linear operations, hence Field Programmable Gate Arrays (FPGA) based hardware acceleration can easily be designed for energy efficient real-time control.
KW - Actor-critic algorithm
KW - Continuous trajectory tracking
KW - Deep reinforcement learning
KW - Unmanned aerial vehicles
UR - http://www.scopus.com/inward/record.url?scp=85074213008&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85074213008&partnerID=8YFLogxK
U2 - 10.1109/RTCSA.2019.8864571
DO - 10.1109/RTCSA.2019.8864571
M3 - Conference contribution
AN - SCOPUS:85074213008
T3 - Proceedings - 2019 IEEE 25th International Conference on Embedded and Real-Time Computing Systems and Applications, RTCSA 2019
BT - Proceedings - 2019 IEEE 25th International Conference on Embedded and Real-Time Computing Systems and Applications, RTCSA 2019
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 25th IEEE International Conference on Embedded and Real-Time Computing Systems and Applications, RTCSA 2019
Y2 - 18 August 2019 through 21 August 2019
ER -