期刊文献+

UAV navigation in high dynamic environments:A deep reinforcement learning approach 被引量:19

原文传递
导出
摘要 Unmanned Aerial Vehicle(UAV) navigation is aimed at guiding a UAV to the desired destinations along a collision-free and efficient path without human interventions, and it plays a crucial role in autonomous missions in harsh environments. The recently emerging Deep Reinforcement Learning(DRL) methods have shown promise for addressing the UAV navigation problem,but most of these methods cannot converge due to the massive amounts of interactive data when a UAV is navigating in high dynamic environments, where there are numerous obstacles moving fast.In this work, we propose an improved DRL-based method to tackle these fundamental limitations.To be specific, we develop a distributed DRL framework to decompose the UAV navigation task into two simpler sub-tasks, each of which is solved through the designed Long Short-Term Memory(LSTM) based DRL network by using only part of the interactive data. Furthermore, a clipped DRL loss function is proposed to closely stack the two sub-solutions into one integral for the UAV navigation problem. Extensive simulation results are provided to corroborate the superiority of the proposed method in terms of the convergence and effectiveness compared with those of the state-of-the-art DRL methods.
出处 《Chinese Journal of Aeronautics》 SCIE EI CAS CSCD 2021年第2期479-489,共11页 中国航空学报(英文版)
基金 supported in part by the National Natural Science Foundation of China (Nos. 61671031, 61722102, and91738301)。
  • 相关文献

参考文献4

二级参考文献10

共引文献113

同被引文献114

引证文献19

二级引证文献74

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部