航空学报 > 2025, Vol. 46 Issue (3): 630553-630553   doi: 10.7527/S1000-6893.2024.30553

深空光电测量与智能感知技术专栏

基于安全强化学习的月球着陆器控制

杨敏, 刘关俊(), 周子渊   

  1. 同济大学 计算机科学与技术系,上海 201804
  • 收稿日期:2024-04-19 修回日期:2024-05-07 接受日期:2024-07-24 出版日期:2024-08-21 发布日期:2024-08-20
  • 通讯作者: 刘关俊 E-mail:liuguanjun@tongji.edu.cn
  • 基金资助:
    国家自然科学基金(62172299);北京控制工程研究所空间光电测量与感知实验室开放基金(LabSOMP-2023-03);中央高校基本科研业务费专项资金(2023-4-YB-05);上海市科技创新行动计划(22511105500)

Control of lunar landers based on secure reinforcement learning

Min YANG, Guanjun LIU(), Ziyuan ZHOU   

  1. Department of Computer Science and Technology,Tongji University,Shanghai 201804,China
  • Received:2024-04-19 Revised:2024-05-07 Accepted:2024-07-24 Online:2024-08-21 Published:2024-08-20
  • Contact: Guanjun LIU E-mail:liuguanjun@tongji.edu.cn
  • Supported by:
    National Natural Science Foundation of China(62172299);Space Optoelectronic Measurement and Perception Lab., Beijing Institute of Control Engineering(LabSOMP-2023-03);The Fundamental Research Funds for the Central Universities(2023-4-YB-05);Shanghai Technological Innovation Action Plan(22511105500)

摘要:

在月球着陆任务中,着陆器必须在极端环境下进行精确操作,并且通常面临着通信延迟的挑战,这些因素严重限制了地面控制的实时操作能力。针对这些挑战,研究提出了一种基于半马尔可夫决策过程(SMDP)的深度强化学习安全性提升框架,旨在提高航天器自主着陆的操作安全性。为了实现状态空间的压缩并保持决策过程的关键特征,该框架通过对历史轨迹的马尔可夫决策过程(MDP)压缩成SMDP,并根据压缩后的轨迹数据构建抽象SMDP状态转移图,然后识别潜在风险的关键状态-动作对,并实施实时监控及干预,有效提高了航天器的自主着陆安全性。采用了反向广度优先搜索方法,搜索出对任务结果有决定性影响的状态-动作对,并通过搭建的状态-动作监控器实现对模型的实时调整。实验结果显示,该框架在不需增加额外传感器或显著改变现有系统配置的条件下,能够在预训练的深度Q网络(DQN)、Dueling DQN、DDQN模型上,提升月球着陆器在模拟环境中的任务成功率高达22%,在预设的安全性评价标准下,该框架能提升最高42%的安全性。此外,虚拟环境中的模拟结果展示了该框架在月球着陆等复杂航天任务中的实际应用潜力,可以有效提升操作安全性和效率。

关键词: 深度强化学习, 自主着陆, 抽象SMDP状态转移图, 安全性提升, 实时监控, 反向广度优先搜索

Abstract:

In lunar landing missions, the lander must perform precise operations in extreme environments and often faces the challenge of communication delays. These factors severely limit the real-time operation capabilities of ground control. In response to these challenges, this study proposes a Deep Reinforcement Learning (DRL) framework for safety enhancement based on the Semi-Markov Decision Process (SMDP) to improve the operational safety of autonomous spacecraft landing. To compress the state space and maintain the key characteristics of the decision-making process, this framework compresses the Markov Decision Process (MDP) of the historical trajectory into a SMDP, and constructs an abstract SMDP state transition diagram based on the compressed trajectory. Then, the key state-action pairs of potential risks are identified, and the real-time monitoring and intervention strategy is implemented. The framework effectively improves the safety of the spacecraft’s autonomous landing. Furthermore, the reverse breadth first search method is used to search for the state-action pairs that have decisive impact on task results, and real-time adjustment of the model is realized through the built state-action monitor. Experimental results show that this framework increases the mission success rate of the lunar lander by up to 22% in a simulated environment on the pre-trained Deep Q-Network (DQN), Dueling DQN, and DDQN models without adding additional sensors or significantly changing the existing system configuration. According to the preset safety evaluation standards, the framework can improve safety by up to 42%. In addition, simulation results in a virtual environment demonstrate the practical application potential of this framework in complex space missions such as lunar landing, which can effectively improve operational safety and efficiency.

Key words: deep reinforcement learning, autonomous landing, abstract SMDP state transition diagram, safety enhancement, real-time monitoring, reverse breadth-first search

中图分类号: