导航

Acta Aeronautica et Astronautica Sinica ›› 2025, Vol. 46 ›› Issue (15): 331354.doi: 10.7527/S1000-6893.2024.31354

• Electronics and Electrical Engineering and Control • Previous Articles    

Distributed UAV formation control with virtual structure guided reinforcement learning

Yu WANG(), Zhipeng XIE, Yongjian TIAN, Guanglei MENG   

  1. School of Automation,Shenyang Aerospace University,Shenyang 110136,China
  • Received:2024-10-08 Revised:2025-01-13 Accepted:2025-02-21 Online:2025-03-11 Published:2025-03-06
  • Contact: Yu WANG E-mail:wangyu@sau.edu.cn
  • Supported by:
    National Natural Science Foundation of China(61906125);Basic Research Funds of Liaoning Provincial Universities(LJ232410143020)

Abstract:

In single decision-making models based on reinforcement learning algorithms, the adaptability is often insufficient when handling complex Unmanned Aerial Vehicle(UAV) formation tasks due to limited autonomous decision-making capabilities. To address this, this paper proposes a distributed decision-making method guided by the virtual structure approach integrated with a deep reinforcement learning algorithm. First, to reduce the difficulty of strategy optimization for reinforcement learning algorithms in diverse task environments, the overall task is functionally decomposed. Local task planning is then implemented for individual task scenarios, such as static obstacles, random obstacles, and communication interference. Multiple decision sub-models are constructed along with the design of the calling process between these models. Next, to enhance guidance, the virtual structure method is integrated with the Soft Actor-Critic(SAC) reinforcement learning algorithm to build a distributed decision-making framework. Through decentralized training of each sub-model, the success rate and flexibility of task execution are significantly improved. Finally, a centralized execution approach is adopted, where environmental changes serve as the triggering condition for the dynamic selection and seamless switching betweeen sub-models. This allows the UAV formation to autonomously adjust its formation according to changes in the task environment, achieving the mission objectives while significantly enhancing the overall adaptability and survivability of the swarm. The effectiveness of the method is validated through simulation experiments in multiple scenarios.

Key words: UAV formation control, complex task environment, deep reinforcement Learning, virtual structure method, distributed decision

CLC Number: