Home > Research > Publications & Outputs > A multi-aircraft co-operative trajectory planni...

Electronic data

  • AEI 2025_Pre-print version

    Accepted author manuscript, 6.14 MB, PDF document

    Available under license: CC BY: Creative Commons Attribution 4.0 International License

Links

Text available via DOI:

View graph of relations

A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning

Research output: Contribution to Journal/MagazineJournal articlepeer-review

E-pub ahead of print

Standard

A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning. / Pang, Bizhao; Xu, Xinting ; Zhang, Mincheng et al.
In: Advanced Engineering Informatics, Vol. 65, 103157, 31.05.2025.

Research output: Contribution to Journal/MagazineJournal articlepeer-review

Harvard

APA

Pang, B., Xu, X., Zhang, M., Alam, S., & Lulli, G. (2025). A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning. Advanced Engineering Informatics, 65, Article 103157. Advance online publication. https://doi.org/10.1016/j.aei.2025.103157

Vancouver

Pang B, Xu X, Zhang M, Alam S, Lulli G. A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning. Advanced Engineering Informatics. 2025 May 31;65:103157. Epub 2025 Feb 3. doi: 10.1016/j.aei.2025.103157

Author

Pang, Bizhao ; Xu, Xinting ; Zhang, Mincheng et al. / A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning. In: Advanced Engineering Informatics. 2025 ; Vol. 65.

Bibtex

@article{0c004b14b76a4816afac2818db87a82e,
title = "A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning",
abstract = "Climate change induces an increased frequency of adverse weather, particularly thunderstorms, posing significant safety and efficiency challenges in en route airspace, especially in oceanic regions with limited air traffic control services. These conditions require multi-aircraft cooperative trajectory planning to avoid both dynamic thunderstorms and other aircraft. Existing literature has typically relied on centralized approaches and single agent principles, which lack coordination and robustness when surrounding aircraft or thunderstorms change paths, leading to scalability issues due to heavy trajectory regeneration needs. To address these gaps, this paper introduces a multi-agent cooperative method for autonomous trajectory planning. The problem is modeled as a Decentralized Markov Decision Process (DEC-MDP) and solved using an Independent Deep Deterministic Policy Gradient (IDDPG) learning framework. A shared actor-critic network is trained using combined experiences from all aircraft to optimize joint behavior. During execution, each aircraft acts independently based on its own observations, with coordination ensured through the shared policy. The model is validated through extensive simulations, including uncertainty analysis, baseline comparisons, and ablation studies. Under known thunderstorm paths, the model achieved a 2 % loss of separation rate, increasing to 4 % with random storm paths.ETA uncertainty analysis demonstrated the model{\textquoteright}s robustness, while baseline comparisons with the Fast Marching Tree and centralized DDPG highlighted its scalability and efficiency. These findings contribute to advancing autonomous aircraft operations. ",
keywords = "Air traffic management, Autonomous trajectory planning, Multi-aircraft coordination, Deep reinforcement learning, Dynamic thunderstorm cells, Climate change",
author = "Bizhao Pang and Xinting Xu and Mincheng Zhang and Sameer Alam and Guglielmo Lulli",
year = "2025",
month = feb,
day = "3",
doi = "10.1016/j.aei.2025.103157",
language = "English",
volume = "65",
journal = "Advanced Engineering Informatics",
issn = "1474-0346",
publisher = "Elsevier",

}

RIS

TY - JOUR

T1 - A multi-aircraft co-operative trajectory planning model under dynamic thunderstorm cells using decentralized deep reinforcement learning

AU - Pang, Bizhao

AU - Xu, Xinting

AU - Zhang, Mincheng

AU - Alam, Sameer

AU - Lulli, Guglielmo

PY - 2025/2/3

Y1 - 2025/2/3

N2 - Climate change induces an increased frequency of adverse weather, particularly thunderstorms, posing significant safety and efficiency challenges in en route airspace, especially in oceanic regions with limited air traffic control services. These conditions require multi-aircraft cooperative trajectory planning to avoid both dynamic thunderstorms and other aircraft. Existing literature has typically relied on centralized approaches and single agent principles, which lack coordination and robustness when surrounding aircraft or thunderstorms change paths, leading to scalability issues due to heavy trajectory regeneration needs. To address these gaps, this paper introduces a multi-agent cooperative method for autonomous trajectory planning. The problem is modeled as a Decentralized Markov Decision Process (DEC-MDP) and solved using an Independent Deep Deterministic Policy Gradient (IDDPG) learning framework. A shared actor-critic network is trained using combined experiences from all aircraft to optimize joint behavior. During execution, each aircraft acts independently based on its own observations, with coordination ensured through the shared policy. The model is validated through extensive simulations, including uncertainty analysis, baseline comparisons, and ablation studies. Under known thunderstorm paths, the model achieved a 2 % loss of separation rate, increasing to 4 % with random storm paths.ETA uncertainty analysis demonstrated the model’s robustness, while baseline comparisons with the Fast Marching Tree and centralized DDPG highlighted its scalability and efficiency. These findings contribute to advancing autonomous aircraft operations.

AB - Climate change induces an increased frequency of adverse weather, particularly thunderstorms, posing significant safety and efficiency challenges in en route airspace, especially in oceanic regions with limited air traffic control services. These conditions require multi-aircraft cooperative trajectory planning to avoid both dynamic thunderstorms and other aircraft. Existing literature has typically relied on centralized approaches and single agent principles, which lack coordination and robustness when surrounding aircraft or thunderstorms change paths, leading to scalability issues due to heavy trajectory regeneration needs. To address these gaps, this paper introduces a multi-agent cooperative method for autonomous trajectory planning. The problem is modeled as a Decentralized Markov Decision Process (DEC-MDP) and solved using an Independent Deep Deterministic Policy Gradient (IDDPG) learning framework. A shared actor-critic network is trained using combined experiences from all aircraft to optimize joint behavior. During execution, each aircraft acts independently based on its own observations, with coordination ensured through the shared policy. The model is validated through extensive simulations, including uncertainty analysis, baseline comparisons, and ablation studies. Under known thunderstorm paths, the model achieved a 2 % loss of separation rate, increasing to 4 % with random storm paths.ETA uncertainty analysis demonstrated the model’s robustness, while baseline comparisons with the Fast Marching Tree and centralized DDPG highlighted its scalability and efficiency. These findings contribute to advancing autonomous aircraft operations.

KW - Air traffic management

KW - Autonomous trajectory planning

KW - Multi-aircraft coordination

KW - Deep reinforcement learning

KW - Dynamic thunderstorm cells

KW - Climate change

U2 - 10.1016/j.aei.2025.103157

DO - 10.1016/j.aei.2025.103157

M3 - Journal article

VL - 65

JO - Advanced Engineering Informatics

JF - Advanced Engineering Informatics

SN - 1474-0346

M1 - 103157

ER -