导航

Acta Aeronautica et Astronautica Sinica ›› 2025, Vol. 46 ›› Issue (23): 631968.doi: 10.7527/S1000-6893.2025.31968

• special column • Previous Articles    

Dynamic brightness reconstruction for UAV visible-infrared fusion object detection

Kui LIU1,2, Hao SUN1,2, Han WU1,2, Kefeng JI1,2(), Gangyao KUANG1,2   

  1. 1.College of Electronic Science and Technology,National University of Defense Technology,Changsha 410073,China
    2.State Key Laboratory of Complex Electromagnetic Environment Effects on Electronics and Information System,National University of Defense Technology,Changsha 410073,China
  • Received:2025-03-12 Revised:2025-03-29 Accepted:2025-05-28 Online:2025-06-10 Published:2025-06-06
  • Contact: Kefeng JI E-mail:jikefeng@nudt.edu.cn
  • Supported by:
    National Natural Science Foundation of China(61971426)

Abstract:

The visible-infrared fusion object detection of Unmanned Aerial Vehicles (UAV) has important application value in military and civilian fields such as disaster rescue, security monitoring, and battlefield reconnaissance. However, under low illumination conditions, existing fusion strategies have several limitations, including ignoring uneven lighting in different areas of the same scene and over -reliance on infrared modalities, which results in the potential rich semantic information of visible images in low illumination conditions. In addition, low light further exacerbates the difficulty of cross modal fusion. To address the above problems, a dynamic brightness reconstruction for UAV visible-infrared fusion object detection method is proposed. Firstly, a Super pixel Dynamic Illumination aware Mask (SDIM) module was designed using prior local illumination information. By simulating the dependence of real scenes on different modalities and introducing superpixel information, the problem of object edge feature loss in existing methods was solved. Secondly, considering the problem of feature degradation in low light visible images, a Low Illumination Image Enhancement (LIIE) module was designed to achieve end-to-end optimization of visible image key semantic adaptive enhancement for detection tasks. Finally, a Multi-Scale Feature Cross-Attention Fusion (MFCF) module was designed to address the fusion conflicts caused by cross modal feature heterogeneity. The module constructs a bimodal feature interaction space through a hierarchical cross attention mechanism and adaptively fuses multi-scale features using a dynamic weight allocation strategy. Based on the typical visible-infrared fusion object detection datasets DroneVehicle and VEDAI, the experimental results verified the effectiveness and robustness of the proposed method in visible-infrared fusion target detection tasks under low illumination conditions. Specifically, compared with existing advanced fusion detection algorithms, the proposed method improved the average accuracy (mAP) by 2.3% and 2.2% respectively while maintaining a low number of parameters, and compared with the widely used single-mode YOLOv8 algorithm, the mAP has increased by up to 12.9%. In addition, cross scene experimental results based on the LIS real low illumination dataset further validated the excellent generalization capability of the proposed method.

Key words: low illumination, UAV, visible-infrared, deep learning, object detection

CLC Number: