Autonomous reentry guidance based on planning-correction hierarchical reinforcement learning

  • PENG Gao-Xiang ,
  • WANG Bo ,
  • LIU Lei ,
  • FAN Hui-Jin
Expand
  • 1. Huazhong University of Science and Technology
    2.

Received date: 2025-06-27

  Revised date: 2025-11-09

  Online published: 2025-11-10

Abstract

To enhance the rapid response capability, mission adaptability, and robustness against significant model deviations during aerospace vehicle reentry, this study proposes an autonomous reentry guidance method based on planning-correction hierarchical reinforcement learning (HRL). Addressing the training instability issues in traditional HRL, a planning-correction hierarchical strategy is introduced to eliminate the dependence of upper-level policy training on lower-level state transition data, establishing a dual-layer guidance framework. In the planning layer, a modular RL policy is employed to plan reference angle-of-attack and bank angle profiles, generating global trajectories according to mission requirements to ensure the framework's adaptability. In the correction layer, high-frequency trajectory corrections under model parameter deviations are performed to mitigate the impact of large parameter deviations. Simulation results demonstrate that the dual-layer guidance strategy can handle larger parameter deviations and improve guidance accuracy under significant uncertainties. Compared to the predictor-corrector guidance algorithm, the proposed strategy exhibits superior mission adaptability and real-time performance, enabling autonomous guidance from arbitrary initial positions and orientations.

Cite this article

PENG Gao-Xiang , WANG Bo , LIU Lei , FAN Hui-Jin . Autonomous reentry guidance based on planning-correction hierarchical reinforcement learning[J]. ACTA AERONAUTICAET ASTRONAUTICA SINICA, 0 : 1 -0 . DOI: 10.7527/S1000-6893.2025.32485

Outlines

/