Uav Path Planning and Obstacle Avoidance Based on Deep Learning and Reinforcement Learning

Authors

  • Zhixi Shi School of Computer Science and Technology,Changsha University of Science & Technology, Changsha, China

DOI:

https://doi.org/10.54097/7g9fac16

Keywords:

Deep learning; Reinforcement learning; Unmanned aerial vehicle; Path planning.

Abstract

With the rapid development of the Internet of Things and low-altitude economy, unmanned aerial vehicles (Uavs) have been widely used in data collection, environmental monitoring and other fields.Traditional path planning algorithms have defects such as insufficient generalization ability and poor real-time performance in complex and dynamic environments. Deep reinforcement learning technology provides a new solution for autonomous decision-making and collaborative operation of unmanned aerial vehicles (Uavs). In this paper, the UAV path planning technology based on deep learning and reinforcement learning is systematically reviewed. In the aspect of single UAV, the improved Q-Learning algorithm realizes fast local replanning through dynamic exploration factor and artificial potential field reward function. DDPG, DySAC, SMLTO and other algorithms improve the safety of obstacle avoidance and trajectory smoothness in three-dimensional dynamic environment. In terms of multiple Uavs, the MRF-DQN algorithm combines preferential experience replay and maximum reward frequency function to improve the efficiency of collaborative task completion. The DPPDQN algorithm realizes efficient task allocation and path optimization in large-scale regions through double pre-segmentation and CNN feature extraction. The experimental results show that the above algorithms are superior to the traditional methods in terms of collision rate, planning delay, energy efficiency and task completion rate, which provides reliable technical support for multi-UAV cooperative operation.

Downloads

Download data is not yet available.

References

[1] Han J C. Multi-UAV collaborative path planning based on reinforcement learning. Xi’an University of Technology, 2025.

[2] Gao Y. Research on Path Planning Algorithm for Multi-UAV Data Collection Based on Reinforcement Learning. Qilu Industrial University, 2025.

[3] Zamoum Y, Baiche K, Benkeddad Y, et al. Modern artificial intelligence technics for unmanned aerial vehicles path planning and control. Bulletin of Electrical Engineering and Informatics, 2025, 14(1): 153-172.

[4] Yin Y, Wang X, Zhou J. Q-Learning-based Multi-UAV Cooperative Path Planning Method. Binggong Xuebao/Acta Armamentarii, 2023, 44(2): 484-495.

[5] Fan H Y. Optimization Method of Autonomous Obstacle Avoidance Path Planning for UAV Based on Deep Reinforcement Learning. New Technology and New Products in China, 2025, (15): 8-10.

[6] Cao X W. Research on UAV Path Planning Method Based on Deep Reinforcement Learning. Sichuan University, 2024.

[7] Tang J. Research on trajectory planning and target tracking strategy for Unmanned Aerial Vehicle in multiple obstacle environment. Southwest Jiaotong University, 2023.

[8] Zhang W J. Research on UAV obstacle detection and path planning algorithm. Xi’an Petroleum University, 2023.

[9] Elfatih N M, Ali E S, Saeed R A. Navigation and Trajectory Planning Techniques for Unmanned Aerial Vehicles Swarm//Artificial Intelligence for Robotics and Autonomous Systems Applications. Cham: Springer International Publishing, 2023: 369-404.

[10] Bayerlein H, Theile M, Caccamo M, et al. multi-UAV path planning for wireless data harvesting with deep reinforcement learning. IEEE Open Journal of the Communications Society, 2021, 2: 1171-1187.

[11] Westheider J, Ruckin J, Popović M. Multi-uav adaptive path planning using deep reinforcement learning //2023 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 2023.

Downloads

Published

27-03-2026

Issue

Section

Articles

How to Cite

Shi, Z. (2026). Uav Path Planning and Obstacle Avoidance Based on Deep Learning and Reinforcement Learning. Frontiers in Computing and Intelligent Systems, 16(1), 178-184. https://doi.org/10.54097/7g9fac16