During the last two decades, Unmanned Aerial Vehicles (UAVs) have been employed for a wide range of civil and public domain applications, as well as in missions to Mars. In complex autonomous exploration scenarios, particularly in GPS-denied environments, the software integrated into the Guidance, Navigation, and Control (GNC) system plays a critical role in ensuring UAV stability and autonomy. To meet these requirements and address the limitations of traditional navigation techniques, the development of Deep Reinforcement Learning (DRL) approaches to support decision-making tasks has gained significant traction in recent years. The goal of the paper is twofold: i) to present a comparison between the traditional Proximal Policy Optimization (PPO) and the augmented PPO with a transformer architecture, ii) to achieve smooth and efficient trajectories by designing a continuous physics informed reward function accounting for the Least Action Principle (LAP). The results demonstrate that PPO achieves significantly improved performance when integrated with the transformer, as well as high efficiency of the trained agent when simulating a specific flight path. This enhancement highlights the potential of transformer-based architectures to more effectively address complex decision-making tasks than traditional DRL methods.

Sopegno L., Cirrincione G., Martini S., Rutherford M.J., Livreri P., Valavanis K.P. (2025). Transformer-Based Physics Informed Proximal Policy Optimization for UAV Autonomous Navigation. In 2025 International Conference on Unmanned Aircraft Systems, ICUAS 2025 (pp. 1094-1099). Institute of Electrical and Electronics Engineers Inc. [10.1109/ICUAS65942.2025.11007786].

Transformer-Based Physics Informed Proximal Policy Optimization for UAV Autonomous Navigation

Sopegno L.;Martini S.;Livreri P.;
2025-01-01

Abstract

During the last two decades, Unmanned Aerial Vehicles (UAVs) have been employed for a wide range of civil and public domain applications, as well as in missions to Mars. In complex autonomous exploration scenarios, particularly in GPS-denied environments, the software integrated into the Guidance, Navigation, and Control (GNC) system plays a critical role in ensuring UAV stability and autonomy. To meet these requirements and address the limitations of traditional navigation techniques, the development of Deep Reinforcement Learning (DRL) approaches to support decision-making tasks has gained significant traction in recent years. The goal of the paper is twofold: i) to present a comparison between the traditional Proximal Policy Optimization (PPO) and the augmented PPO with a transformer architecture, ii) to achieve smooth and efficient trajectories by designing a continuous physics informed reward function accounting for the Least Action Principle (LAP). The results demonstrate that PPO achieves significantly improved performance when integrated with the transformer, as well as high efficiency of the trained agent when simulating a specific flight path. This enhancement highlights the potential of transformer-based architectures to more effectively address complex decision-making tasks than traditional DRL methods.
2025
9798331513283
Sopegno L., Cirrincione G., Martini S., Rutherford M.J., Livreri P., Valavanis K.P. (2025). Transformer-Based Physics Informed Proximal Policy Optimization for UAV Autonomous Navigation. In 2025 International Conference on Unmanned Aircraft Systems, ICUAS 2025 (pp. 1094-1099). Institute of Electrical and Electronics Engineers Inc. [10.1109/ICUAS65942.2025.11007786].
File in questo prodotto:
File Dimensione Formato  
Transformer-Based_Physics_Informed_Proximal_Policy_Optimization_for_UAV_Autonomous_Navigation.pdf

Solo gestori archvio

Tipologia: Versione Editoriale
Dimensione 450.31 kB
Formato Adobe PDF
450.31 kB Adobe PDF   Visualizza/Apri   Richiedi una copia

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/10447/683703
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? ND
social impact