TY - GEN
T1 - Neural Network Pruning and Fast Training for DRL-based UAV Trajectory Planning
AU - Li, Yilan
AU - Fang, Haowen
AU - Li, Mingyang
AU - Ma, Yue
AU - Qiu, Qinru
N1 - Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - Deep reinforcement learning (DRL) has been applied for optimal control of autonomous UAV trajectory generation. The energy and payload capacity of small UAVs impose constraints on the complexity and size of the neural network. While Model compression has the potential to optimize the trained neural network model for efficient deployment on em-bedded platforms, pruning a neural network for DRL is more difficult due to the slow convergence in the training before and after pruning. In this work, we focus on improving the speed of DRL training and pruning. New reward function and action exploration are first introduced, resulting in convergence speedup by 34.14%. The framework that integrates pruning and DRL training is then presented with an emphasize on how to reduce the training cost. The pruning does not only improve computational performance of inference, but also reduces the training effort with-out compromising the quality of the trajectory. Finally, experimental results are presented. We show that the integrated training and pruning framework reduces 67.16% of the weight and improves trajectory success rate by 1.7%. It achieves a 4.43x reduction of the floating-point operations for the inference, resulting a measured 41.85% run time reduction.
AB - Deep reinforcement learning (DRL) has been applied for optimal control of autonomous UAV trajectory generation. The energy and payload capacity of small UAVs impose constraints on the complexity and size of the neural network. While Model compression has the potential to optimize the trained neural network model for efficient deployment on em-bedded platforms, pruning a neural network for DRL is more difficult due to the slow convergence in the training before and after pruning. In this work, we focus on improving the speed of DRL training and pruning. New reward function and action exploration are first introduced, resulting in convergence speedup by 34.14%. The framework that integrates pruning and DRL training is then presented with an emphasize on how to reduce the training cost. The pruning does not only improve computational performance of inference, but also reduces the training effort with-out compromising the quality of the trajectory. Finally, experimental results are presented. We show that the integrated training and pruning framework reduces 67.16% of the weight and improves trajectory success rate by 1.7%. It achieves a 4.43x reduction of the floating-point operations for the inference, resulting a measured 41.85% run time reduction.
UR - http://www.scopus.com/inward/record.url?scp=85126131849&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85126131849&partnerID=8YFLogxK
U2 - 10.1109/ASP-DAC52403.2022.9712561
DO - 10.1109/ASP-DAC52403.2022.9712561
M3 - Conference contribution
AN - SCOPUS:85126131849
T3 - Proceedings of the Asia and South Pacific Design Automation Conference, ASP-DAC
SP - 574
EP - 579
BT - ASP-DAC 2022 - 27th Asia and South Pacific Design Automation Conference, Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 27th Asia and South Pacific Design Automation Conference, ASP-DAC 2022
Y2 - 17 January 2022 through 20 January 2022
ER -