ACTA AERONAUTICAET ASTRONAUTICA SINICA >
High-precision monocular vision pose measurement for large distance span in carrier landing guidance
Received date: 2024-11-25
Revised date: 2024-12-17
Accepted date: 2025-01-20
Online published: 2025-02-21
Supported by
National Natural Science Foundation of China(12272404)
The autonomous landing guidance involves a large distance span, resulting in significant scale variations of the ship target in the image sequences obtained through monocular vision guidance. Existing pose measurement methods struggle to achieve high-precision monocular vision pose measurement across such a wide distance range. For current monocular vision pose measurement methods based on sparse keypoint sets, this paper focuses on improving the accuracy of keypoint detection, and analyzes the impact of target size and network input size on keypoint detection accuracy. Furthermore, this paper proposes a novel monocular vision pose measurement method based on multiple components, balancing both accuracy and efficiency. By using sparse keypoint sets to represent components in a simplified manner, and building on a coarse pose estimation of the overall ship target components, this method introduces a path aggregation feature pyramid network and a hierarchical encoding module to achieve high-precision detection of local component keypoints. Subsequently, by integrating the high-precision keypoint detection results of all components and solving the Perspective-n-Points (PnP) problem, the method achieves robust and high-precision pose measurement across the large distance span required for landing guidance. Simulation experiments and scaled physical experiments demonstrate that the proposed method achieves robust and high-precision monocular pose measurement across the large distance span for landing guidance, outperforming existing methods, with an average single-frame inference time of approximately 40 ms on embedded platforms.
Key words: monocular; landing guidance; pose measurement; deep learning; keypoint detection
Lin CHEN , Xiwen GU , Zhiying CHEN , Zhuo ZHANG , Xiaoliang SUN . High-precision monocular vision pose measurement for large distance span in carrier landing guidance[J]. ACTA AERONAUTICAET ASTRONAUTICA SINICA, 2025 , 46(15) : 331568 -331568 . DOI: 10.7527/S1000-6893.2025.31568
| [1] | WICKRAMASURIYA M, LEE T, SNYDER M. Deep monocular relative 6D pose estimation for ship-based autonomous UAV?[C]?∥AIAA SCITECH 2024 Forum. Reston: AIAA, 2024. |
| [2] | 胡小兵, 周大鹏, 曲晓雷. 国外舰载机全自动着舰技术综述[J]. 飞机设计, 2021, 41(2): 32-36. |
| HU X B, ZHOU D P, QU X L. Review on full automatic carrier landing technique of foreign shipboard aircraft?[J]. Aircraft Design, 2021, 41(2): 32-36 (in Chinese). | |
| [3] | 甄子洋, 王新华, 江驹, 等. 舰载机自动着舰引导与控制研究进展[J]. 航空学报, 2017, 38(2): 020435. |
| ZHEN Z Y, WANG X H, JIANG J, et al. Research progress in guidance and control of automatic carrier landing of carrier-based aircraft[J]. Acta Aeronautica et Astronautica Sinica, 2017, 38(2): 020435 (in Chinese). | |
| [4] | 张志冰, 甄子洋, 江驹, 等. 舰载机自动着舰引导与控制综述[J]. 南京航空航天大学学报, 2018, 50(6): 734-744. |
| ZHANG Z B, ZHEN Z Y, JIANG J, et al. Review on development in guidance and control of automatic carrier landing of carrier-based aircraft[J]. Journal of Nanjing University of Aeronautics & Astronautics, 2018, 50(6): 734-744 (in Chinese). | |
| [5] | 魏振忠. 舰载机着舰位姿视觉测量技术概述[J]. 测控技术, 2020, 39(8): 2-6. |
| WEI Z Z. Overview of visual measurement technology for landing position and attitude of carrier-based aircraft[J]. Measurement & Control Technology, 2020, 39(8): 2-6 (in Chinese). | |
| [6] | GUI Y, GUO P Y, ZHANG H L, et al. Airborne vision-based navigation method for UAV accuracy landing using infrared lamps[J]. Journal of Intelligent & Robotic Systems, 2013, 72(2): 197-218. |
| [7] | POLVARA R, SHARMA S, WAN J, et al. Towards autonomous landing on a moving vessel through fiducial markers[C]∥2017 European Conference on Mobile Robots (ECMR). Piscataway: IEEE Press, 2017: 1-6. |
| [8] | XU G L, QI X P, ZENG Q H, et al. Use of Land’s cooperative object to estimate UAV’s pose for autonomous landing[J]. Chinese Journal of Aeronautics, 2013, 26(6): 1498-1505. |
| [9] | SHI F H, ZHANG X Y, LIU Y C. A new method of camera pose estimation using 2D-3D corner correspondence[J]. Pattern Recognition Letters, 2004, 25(10): 1155-1163. |
| [10] | ZHOU L P, KOPPEL D, KAESS M. A complete, accurate and efficient solution for the perspective-N-line problem[J]. IEEE Robotics and Automation Letters, 2021, 6(2): 699-706. |
| [11] | 毕道明, 黄辉, 范静, 等. 视觉着舰中非合作结构化特征匹配算法[J]. 南京航空航天大学学报, 2021, 53(3): 395-401. |
| BI D M, HUANG H, FAN J, et al. Non-cooperative structural feature matching algorithm in visual landing[J]. Journal of Nanjing University of Aeronautics & Astronautics, 2021, 53(3): 395-401 (in Chinese). | |
| [12] | 王秋富, 石治国, 张倬, 等. 舰载机着舰引导中鲁棒单目视觉相对位姿测量[J]. 航空学报, 2024, 45 (23): 330309. |
| WANG Q F, SHI Z G, ZHANG Z, et al. Robust monocular relative pose measurement for carrier-based aircraft landing guidance[J]. Acta Aeronautica et Astronautica Sinica, 2024, 45 (23): 330309 (in Chinese). | |
| [13] | SUN X L, ZHANG Z, LIU J, et al. Visual pose measurement for automatic landing on an aircraft carrier[C]∥2022 IEEE International Conference on Unmanned Systems (ICUS). Piscataway: IEEE Press, 2022: 891-895. |
| [14] | ZHANG Z, WANG Q F, BI D M, et al. MC-LRF based pose measurement system for shipborne aircraft automatic landing?[J]. Chinese Journal of Aeronautics, 2023, 36(8): 298-312. |
| [15] | WICKRAMASURIYA M, LEE T, SNYDER M. Deep transformer network for monocular pose estimation of ship-based UAV[DB/OL]. arXiv preprint: 2406.09260, 2024. |
| [16] | YU G H, CHANG Q Y, LV W Y, et al. PP-PicoDet: A better real-time object detector on mobile devices[DB/OL]. arXiv preprint: 2111.00902, 2021. |
| [17] | JIANG T, LU P, ZHANG L, et al. RTMPose: real-time multi-person pose estimation based on MMPose[DB/OL]. arXiv preprint: 2303.07399, 2023. |
| [18] | FISCHLER M A, BOLLES R C. Random sample consensus[J]. Communications of the ACM, 1981, 24(6): 381-395. |
| [19] | LEPETIT V, MORENO-NOGUER F, FUA P. EPnP: An accurate O(n) solution to the PnP problem[J]. International Journal of Computer Vision, 2009, 81(2): 155-166. |
| [20] | LIU S, QI L, QIN H F, et al. Path aggregation network for instance segmentation[C]∥2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Piscataway: IEEE Press, 2018: 8759-8768. |
| [21] | RAZAVI A, VAN DEN OORD A, VINYALS O. Generating diverse high-fidelity images with VQ-VAE-2[DB/OL]. arXiv preprint: 1906.00446, 2019. |
| [22] | LI Y J, YANG S, LIU P D, et al. SimCC: A simple coordinate classification perspective for human pose estimation?[M]?∥Computer Vision-ECCV 2022. Cham: Springer Nature Switzerland, 2022: 89-106. |
| [23] | GitHub. MMPose: OpenMMlab pose estimation toolbox and benchmark[DB/OL]. . |
| [24] | PaddlePaddle.? ?Object detection and instance segmentation toolkit based on PaddlePaddle[DB/OL]. (2024-12-05)[2025-01-08]. . |
| [25] | XU Y F, ZHANG J, ZHANG Q M, et al. ViTPose: Simple vision transformer baselines for human pose estimation[J]. Advances in Neural Information Processing Systems, 2022, 35: 38571-38584. |
| [26] | GUO L C, CHEN L, WANG Q F, et al. Joint optimization of the 3D model and 6D pose for monocular pose estimation[J]. Drones, 2024, 8(11): 626. |
| [27] | DENNINGER M, SUNDERMEYER M, WINKELBAUER D, et al. BlenderProc?[DB/OL]. arXiv preprint: 1911.01911, 2019. |
/
| 〈 |
|
〉 |