电子电气工程与控制

GPS拒止环境巡飞器双源地理位置知觉实例分割

  • 宫冕 ,
  • 张印辉 ,
  • 何自芬 ,
  • 陈光晨 ,
  • 张瑞
展开
  • 昆明理工大学 机电工程学院,昆明 650500

收稿日期: 2025-04-15

  修回日期: 2025-04-24

  录用日期: 2025-06-06

  网络出版日期: 2025-06-16

基金资助

国家自然科学基金(62061022);国家自然科学基金(62171206);装备智能运用教育部重点实验室开放基金(AAIE-2023-0203);昆明理工大学分析测试基金(2024P20231103003)

Instance segmentation of dual-source geolocation perception for GPS reject environmental aerial vehicle

  • Mian GONG ,
  • Yinhui ZHANG ,
  • Zifen HE ,
  • Guangchen CHEN ,
  • Rui ZHANG
Expand
  • Mechanical and Electrical Engineering,Kunming University of Science and Technology,Kunming 650500,China

Received date: 2025-04-15

  Revised date: 2025-04-24

  Accepted date: 2025-06-06

  Online published: 2025-06-16

Supported by

National Natural Science Foundation of China(62061022);The Open Fund Project of the Key Laboratory of Intelligent Application of Equipment of the Ministry of Education(AAIE-2023-0203);Analysis and Testing Foundation of Kunming University of Science and Technology(2024P20231103003)

摘要

针对巡飞器导航系统GPS拒止环境下无法定位及其航拍图像地物目标与背景混杂且目标边缘模糊导致分割精度低的问题,提出一种双源地理位置知觉地物目标实例分割模型(DGISM)。首先,设计沿水平、竖直和通道维度的三重自适应中心矩模块同时引入并行路径跨层交互结构聚合浅层细节信息,以提高实例特征区分度从而克服背景混杂问题。其次,设计多核空间聚焦模块融合跨尺度空间特征信息和深层通道信息以增强模型对边缘特征的聚集能力,改善了因边缘模糊导致分割精度低的问题。最后,针对GPS拒止环境下无法定位的问题,设计航拍图像与卫星信息双源感知定位模块,经分割权重提取巡飞器航拍图像特征信息与具有卫星定位信息的图像特征记忆库相匹配获取对应地理位置信息。实验结果表明:提出的双源地理位置知觉地物目标实例分割模型mAP50-95和mAP50分割精度分别达到69.1%和88.8%,图像特征双源匹配定位精度达到95.6%,在GPS拒止环境下通过分割建筑类地物目标实现巡飞器地理位置知觉精确定位。

本文引用格式

宫冕 , 张印辉 , 何自芬 , 陈光晨 , 张瑞 . GPS拒止环境巡飞器双源地理位置知觉实例分割[J]. 航空学报, 2026 , 47(2) : 332120 -332120 . DOI: 10.7527/S1000-6893.2025.32120

Abstract

To address the issues of inability to locate in GPS reject environments for the navigation system of aerial vehicles and the low seg-mentation accuracy caused by blurred edges and mixed backgrounds in aerial imagery, this paper proposes a Dual-source Geo-aware Instance Segmentation Model for ground objects, referred to as DGISM. First, a triple adaptive central moment module along horizontal, vertical, and channel dimensions is designed, along with a parallel path cross-layer interaction structure to aggregate shallow-level details, thereby enhancing instance feature discriminability and overcoming background clutter issues. Second, a multi-core spatial focus module is developed to integrate cross-scale spatial features and deep channel information, improving the model’s ability to concentrate on edge features and addressing the problem of low segmentation accuracy due to edge blur. Lastly, to tackle the problem of localization in GPS reject environments, a dual-source perception and positioning module for aerial imagery and satellite information is designed. This module extracts segmentation weights to match the feature information of the aerial vehicle’s imagery with a feature memory bank of images containing satellite positioning information, thereby obtaining corresponding geographic location information. Experimental results demonstrate that the proposed dual-source geo-aware instance segmentation model achieves mAP50-95 and mAP50 segmentation accuracies of 69.1% and 88.8%, respectively, with a dual-source feature matching positioning accuracy of 95.6%. Realizing precise geographic location perception and positioning of patrol aircraft by segmenting building targets in GPS reject environments

参考文献

[1] 屈若锟, 王致远, 刘晔璐, 等 .面向城市空中交通的无人机视觉定位技术研究[J].航空学报202546(10): 531168.
  QU R K, WANG ZY, LIU Y L, et al. Research on UAV visual positioning technology for urban air mobility[J]. Acta Aeronautica et Astronautica Sinica202546(10): 531168 (in Chinese)
[2] HE K, GKIOXARI G, DOLLáR P, et al. Mask R-CNN[C]∥Proceedings of the IEEE International Conference on Computer Vision. Piscataway: IEEE Press, 2017: 2961-2969.
[3] CHEN K, PANG J, WANG J, et al. Hybrid task cascade for instance segmentation[C]∥Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Piscataway: IEEE Press, 2019: 4974-4983.
[4] CHENG B, MISRA I, SCHWING A G, et al. Masked-attention mask transformer for universal image segmentation[C]∥Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Piscataway: IEEE Press, 2022: 1290-1299.
[5] CHEN X, GIRSHICK R, He K, et al. Tensormask: A foundation for dense object segmentation[C]∥Proceedings of the IEEE/CVF International Conference on Computer vision. Piscataway: IEEE Press, 2019: 2061-2069.
[6] REDMON J, DIVVALA S, GIRSHICK R, et al. You only look once: Unified, real-time object detection[C]∥Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Piscataway: IEEE Press, 2016: 779-788.
[7] WANG X, ZHANG R, KONG T, et al. Solov2: Dynamic and fast instance segmentation[J]. Advances in Neural Information Processing Systems202033: 17721-17732.
[8] ZHOU C. Yolact++ better real-time instance segmentation[M]. California: University of California, 2020: 1108-1121.
[9] BOLYA D, ZHOU C, XIAO F, et al. Yolact: Real-time instance segmentation[C]∥Proceedings of the IEEE/CVF International Conference on Computer Vision. Piscataway: IEEE Press,2019: 9157-9166.
[10] DONG C, TANG Y, ZHANG L. GHA-Inst: A real-time instance segmentation model utilizing YOLO detection framework[J]. Cluster Computing202427(6): 7401-7415.
[11] WANG C Y, BOCHKOVSKIY A, LIAO H Y M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors[C]∥Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Piscataway: IEEE Press, 2023: 7464-7475.
[12] 姚俞成,李旭,徐启敏,等 .基于特征增强和校准的航拍车辆实例分割方法[J].航空学报202344(24): 328397.
  YAO Y C, LI X, XU Q M, et al. Instance segmentation for vehicle in UAV aerial images based on feature enhancement and calibration[J]. Acta Aeronautica et Astronautica Sinica202344(24): 328397 (in Chinese).
[13] CHEN Y, YUAN X, WANG J, et al. YOLO-MS: Rethinking multi-scale representation learning for real-time object detection[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence202547(6): 4240-4252.
[14] WU H, ZHANG M, HUANG P, et al. CMLFormer: CNN and multi-scale local-context transformer network for remote sensing images semantic segmentation[J]. IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing202417: 7233-7241.
[15] CHEN N H, LOU T S, ZHAO L Y. MMHCA: Multi-feature representations based on multi-scale hierarchical contextual aggregation for UAV-view geo-localization[J]. Chinese Journal of Aeronautics202538(6): 103242.
[16] 曹正阳, 张冰, 白屹轩, 等. GNSS/INS/VNS组合定位信息融合的多无人机协同导航方法[J].兵工学报202344(S2): 157-166.
  CAO Z Y, ZHANG B, BAI Y X, et al. GNSS/INS/VNS integrated positioning information fusion method for multi UAV cooperative navigation[J]. Acta Armamentarii202344(S2): 157-166 (in Chinese)
[17] SUN K, MOHTA K, PFROMMER B, et al. Robust stereo visual inertial odometry for fast autonomous flight[J]. IEEE Robotics and Automation Letters20183(2): 965-972.
[18] NASSAR A, AMER K, ELHAKIM R, et al. A deep CNN-based framework for enhanced aerial imagery registration with applications to UAV geolocalization[C]∥Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops. Piscataway: IEEE Press, 2018: 1513-1523.
[19] CHOI J, MYUNG H. BRM localization: UAV localization in GNSS-denied environments based on matching of numerical map and UAV images[C]∥2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). Piscataway: IEEE Press, 2020: 4537-4544.
[20] Ultralytics. YOLOv8 2023 [EB/OL]. .
[21] 邓立暖,尧新峰. 基于NSST的红外与可见光图像融合算法[J]. 电子学报201745(12): 2965-2970.
  DENG L Y, YAO X F, et al [J]. Infrared and visible image fusion algorithm based on NSST. Journal of Electronics201745(12): 2965-2970 (in Chinese)
[22] DAI M, ZHENG E, FENG Z, et al. Vision-based UAV self-positioning in low-altitude urban environments[J]. IEEE Transactions on Image Processing202333: 493-508.
[23] HU J, SHEN L, SUN G. Squeeze-and-excitation networks[C]∥Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Piscataway: IEEE Press, 2018: 7132-7141.
[24] ROY A G, NAVAB N, WACHINGER C. Concurrent spatial and channel ‘squeeze & excitation’in fully convolutional networks[C]∥Medical Image Computing and Computer Assisted Intervention-MICCAI 2018: 21st International Conference, 2018: 421-429.
[25] CAO Y, XU J, LIN S, et al. Gcnet: Non-local networks meet squeeze-excitation networks and beyond[C]∥Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops. Piscataway: IEEE Press, 2019.
[26] ZHANG P, RAN H, JIA C, et al. A lightweight propagation path aggregating network with neural topic model for rumor detection[J]. Neurocomputing2021458: 468-477.
[27] ZHANG X, ZENG H, GUO S, et al. Efficient long-range attention network for image super-resolution[C]∥European Conference on Computer Vision. Cham: Springer Nature Switzerland, 2022: 649-667.
[28] HE K, ZHANG X, REN S, et al. Deep residual learning for image recognition[C]∥Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Piscataway: IEEE Press, 2016: 770-778.
[29] LIU X, PENG H, ZHENG N, et al. Efficientvit: Memory efficient vision transformer with cascaded group attention[C]∥Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Piscataway: IEEE Press, 2023: 14420-14430.
[30] LIU Z, LIN Y, CAO Y, et al. Swin transformer: Hierarchical vision transformer using shifted windows[C]∥Proceedings of the IEEE/CVF International Conference on Computer Vision. Piscataway: IEEE Press, 2021: 10012-10022.
[31] MEHTA S, RASTEGARI M. Mobilevit: Light-weight, general-purpose, and mobile-friendly vision transformer[DB/OL]. arXiv preprint:2110.02178, 2021.
[32] CHEN J, KAO S, HE H, et al. Run, don’t walk: Chasing higher FLOPS for faster neural networks[C]∥Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Piscataway: IEEE Press, 2023: 12021-12031.
文章导航

/