No CrossRef data available.
Published online by Cambridge University Press: 17 November 2025
Reinforcement learning (RL) has demonstrated computational efficiency and autonomy in solving unmanned aerial vehicle (UAV) obstacle avoidance problems. However, practical applications still remain challenges, such as three-dimensional manoeuvres, dynamic obstacles and kinematic constraints. This paper proposes a real-time obstacle avoidance method based on RL and a kinematic model, where the RL framework outputs 3D-axis velocity to enable continuous UAV manoeuver control. To perceive large-scale, highly dynamic obstacles, we establish a 3D safety separation model and construct a modular observation matrix to enhance perception capability. The Soft Actor-Critic (SAC) algorithm is adopted to enhance stochastic exploration in high-dimensional state space while incorporating flight uncertainty. Under simulation, the proposed method effectively avoids dynamic obstacles. The optimised state space boosts learning speed and performance. This provides an effective solution for UAV autonomous obstacle avoidance in complex environments.