A Deep Reinforcement Learning-Based Adaptive Control Strategy for UAVs in Dynamic and Complex Environments

Authors

  • Baraa M. Albaker Department of Electrical Engineering, Collage of Engineering, Al-Iraqia University, Baghdad, Iraq

DOI:

https://doi.org/10.58564/IJSER.4.1.2025.295

Keywords:

Adaptive Control, Collision Avoidance, Deep Reinforcement Learning, Unmanned Aerial Vehicles, UAV navigation

Abstract

The increasing usage of Unmanned Aerial Vehicles (UAVs) in diverse applications necessitates the development of effective flight control systems. A major difficulty, nevertheless, is maintaining robust flight control under complex and dynamic environmental conditions, such as obstacle and non-flying zones, wind disturbances, and sensor noise. Traditional control techniques fail to achieve required flight control in these environments. To address this point, an adaptive control strategy is proposed based on a Deep Reinforcement Learning (DRL) model to enhance the flight performance of quadcopters. The learning and adaptation of the DRL-based control strategy are implemented in real-time through continuous interaction with the environment. This is to improve flight control and achieve consistent UAV performance under varying conditions. Proximal policy optimization with a reward function is used to minimize positional errors, ensure collision-free flight paths, and reduce energy consumption. The developed DRL model for quadcopters is trained in a simulated environment and then tested in three complex environmental scenarios, including urban, forest and mountain terrains. Experimental results demonstrate remarkable improvements in UAV flight performance. In the training phase, the reported training reward increased from 10 to 110 and the train loss is dropped from 0.85 to 0.05, which indicated successful model learning. Also, during system verification, rewards increased from 12 to 115 and UAV flight path deviations were decreased from 0.5 to 0.08m. The proposed controller outperforms conventional approaches in urban environments by lowering average trajectory deviations to 0.2m from 0.35m for MPC and 0.6m for PID. Also, the developed DRL-based controller outperformed the PID and MPC controllers, with path deviations of 0.18m in mountains and 0.12m in forests. In addition, fewer collision rates with obstacles are achieved with the model, 3% in forest, 1.8% in urban areas, and 4.5% in mountains. Furthermore, the consumed energy is reduced to 950J as compared to 1200J for PID and 1050J for MPC. The results show the strength of deploying the proposed controller in following the intended UAV flight path with high precision, effectively avoiding detected conflicts and minimizing consumed energy by the UAV.

References

[1] S. A. H. Mohsan, M. A. Khan, F. Noor, I. Ullah, and M. H. Alsharif, “Towards the unmanned aerial vehicles (UAVs): A comprehensive review,” Drones, vol. 6, no. 6, p. 147, 2022.

[2] H. Shakhatreh et al., “Unmanned aerial vehicles (UAVs): A survey on civil applications and key research challenges,” Ieee Access, vol. 7, pp. 48572–48634, 2019.

[3] S. A. H. Mohsan, N. Q. H. Othman, Y. Li, M. H. Alsharif, and M. A. Khan, “Unmanned aerial vehicles (UAVs): Practical aspects, applications, open challenges, security issues, and future trends,” Intell Serv Robot, vol. 16, no. 1, pp. 109–137, 2023.

[4] Z. Zuo, C. Liu, Q.-L. Han, and J. Song, “Unmanned aerial vehicles: Control methods and future challenges,” IEEE/CAA Journal of Automatica Sinica, vol. 9, no. 4, pp. 601–614, 2022.

[5] B. M. A. Albaker and N. A. R. Rahim, “A Conceptual Framework and a Review of Conflict Sensing, Detection, Awareness and Escape Maneuvering Methods for UAVs,” in Aeronautics and astronautics, IntechOpen, 2011.

[6] N. Bashir, S. Boudjit, G. Dauphin, and S. Zeadally, “An obstacle avoidance approach for UAV path planning,” Simul Model Pract Theory, vol. 129, p. 102815, 2023.

[7] L. J. Mpanza and J. O. Pedro, “Optimised tuning of a PID-Based flight controller for a medium-scale rotorcraft,” Algorithms, vol. 14, no. 6, p. 178, 2021.

[8] B. M. Albaker and N. A. Rahim, “Flight path PID controller for propeller-driven fixed-wing unmanned aerial vehicles,” International Journal of Physical Sciences, vol. 6, no. 8, 2011.

[9] L. Bauersfeld, L. Spannagl, G. J. J. Ducard, and C. H. Onder, “MPC flight control for a tilt-rotor VTOL aircraft,” IEEE Trans Aerosp Electron Syst, vol. 57, no. 4, pp. 2395–2409, 2021.

[10] M. E.-S. M. Essa, M. Elsisi, M. Saleh Elsayed, M. Fawzy Ahmed, and A. M. Elshafeey, “An improvement of model predictive for aircraft longitudinal flight control based on intelligent technique,” Mathematics, vol. 10, no. 19, p. 3510, 2022.

[11] S. Sun, A. Romero, P. Foehn, E. Kaufmann, and D. Scaramuzza, “A comparative study of nonlinear mpc and differential-flatness-based control for quadrotor agile flight,” IEEE Transactions on Robotics, vol. 38, no. 6, pp. 3357–3373, 2022.

[12] T. Susanto, M. B. Setiawan, A. Jayadi, F. Rossi, A. Hamdhi, and J. P. Sembiring, “Application of Unmanned Aircraft PID Control System for Roll, Pitch and Yaw Stability on Fixed Wings,” in 2021 International Conference on Computer Science, Information Technology, and Electrical Engineering (ICOMITEE), IEEE, 2021, pp. 186–190.

[13] M. A. Kamel, A. T. Hafez, and X. Yu, “A review on motion control of unmanned ground and aerial vehicles based on model predictive control techniques,” Journal of Engineering Science and Military Technologies, vol. 2, no. 1, pp. 10–23, 2018.

[14] B. Ye, J. Li, J. Li, C. Liu, J. Li, and Y. Yang, “Deep reinforcement learning-based diving/pull-out control for bioinspired morphing UAVs,” Unmanned Systems, vol. 11, no. 02, pp. 191–202, 2023.

[15] B. Li, W. Zhou, J. Sun, C.-Y. Wen, and C.-K. Chen, “Development of model predictive controller for a Tail-Sitter VTOL UAV in hover flight,” Sensors, vol. 18, no. 9, p. 2859, 2018.

[16] M. Coppola, K. N. McGuire, C. De Wagter, and G. C. H. E. De Croon, “A survey on swarming with micro air vehicles: Fundamental challenges and constraints,” Front Robot AI, vol. 7, p. 18, 2020.

[17] W. Koch, R. Mancuso, R. West, and A. Bestavros, “Reinforcement learning for UAV attitude control,” ACM Transactions on Cyber-Physical Systems, vol. 3, no. 2, pp. 1–21, 2019.

[18] S. Y. Choi and D. Cha, “Unmanned aerial vehicles using machine learning for autonomous flight; state-of-the-art,” Advanced Robotics, vol. 33, no. 6, pp. 265–277, Mar. 2019, doi: 10.1080/01691864.2019.1586760.

[19] K. Wan, X. Gao, Z. Hu, and G. Wu, “Robust motion control for UAV in dynamic uncertain environments using deep reinforcement learning,” Remote Sens (Basel), vol. 12, no. 4, p. 640, 2020.

[20] C. Wang, J. Wang, Y. Shen, and X. Zhang, “Autonomous navigation of UAVs in large-scale complex environments: A deep reinforcement learning approach,” IEEE Trans Veh Technol, vol. 68, no. 3, pp. 2124–2136, 2019.

[21] F. AlMahamid and K. Grolinger, “Autonomous unmanned aerial vehicle navigation using reinforcement learning: A systematic review,” Eng Appl Artif Intell, vol. 115, p. 105321, 2022.

[22] C. Tang, B. Abbatematteo, J. Hu, R. Chandra, R. Martín-Martín, and P. Stone, “Deep reinforcement learning for robotics: A survey of real-world successes,” Annu Rev Control Robot Auton Syst, vol. 8, 2024.

[23] K. Kumar and N. Kumar, “Region coverage-aware path planning for unmanned aerial vehicles: A systematic review,” Physical Communication, vol. 59, p. 102073, 2023.

[24] J. K. Viswanadhapalli, V. K. Elumalai, S. Shivram, S. Shah, and D. Mahajan, “Deep reinforcement learning with reward shaping for tracking control and vibration suppression of flexible link manipulator,” Appl Soft Comput, vol. 152, p. 110756, 2024.

[25] A. Koubâa, A. Allouch, M. Alajlan, Y. Javed, A. Belghith, and M. Khalgui, “Micro air vehicle link (mavlink) in a nutshell: A survey,” IEEE Access, vol. 7, pp. 87658–87680, 2019.

[26] A. Singla, S. Padakandla, and S. Bhatnagar, “Memory-based deep reinforcement learning for obstacle avoidance in UAV with limited environment knowledge,” IEEE transactions on intelligent transportation systems, vol. 22, no. 1, pp. 107–118, 2019.

[27] S. F. Abedin, M. S. Munir, N. H. Tran, Z. Han, and C. S. Hong, “Data freshness and energy-efficient UAV navigation optimization: A deep reinforcement learning approach,” IEEE Transactions on Intelligent Transportation Systems, vol. 22, no. 9, pp. 5994–6006, 2020.

[28] M. Nikolaiev and M. Novotarskyi, “Comparative Review of Drone Simulators,” Information, Computing and Intelligent systems, no. 4, pp. 79–99, 2024.

[29] D. Tejero-Ruiz, F. J. Pérez-Gran, A. Viguria, and A. Ollero, “Realistic Unmanned Aerial Vehicle Simulation: A Comprehensive Approach,” in 2024 7th Iberian Robotics Conference (ROBOT), IEEE, 2024, pp. 1–6.

Downloads

Published

2025-03-01

How to Cite

M. Albaker, B. (2025). A Deep Reinforcement Learning-Based Adaptive Control Strategy for UAVs in Dynamic and Complex Environments . Al-Iraqia Journal for Scientific Engineering Research, 4(1), 77–88. https://doi.org/10.58564/IJSER.4.1.2025.295

Issue

Section

Articles

Most read articles by the same author(s)