Landmark Selection Method for Autonomous Optical Navigation Based on Depth Estimation Error

LI Jiaxing1, WANG Dayi1, DONG Tianshu1, LI Maodeng2, XU Chao2

PDF(1342 KB)
PDF(1342 KB)
Journal of Deep Space Exploration ›› 2023, Vol. 10 ›› Issue (2) : 133-139. DOI: 10.15982/j.issn.2096-9287.2023.20230019
Topic: Celestial Navigation Technology for Deep Space Exploration

Landmark Selection Method for Autonomous Optical Navigation Based on Depth Estimation Error

  • LI Jiaxing1, WANG Dayi1, DONG Tianshu1, LI Maodeng2, XU Chao2
Author information +
History +

Abstract

When autonomous optical navigation is used to achieve accurate planetary landing,the amount of computation required for landmark selection needs to be reduced due to the limitation of computational resources. In this paper,a depth estimation error-based landmark selection method was proposed. First,a depth estimation error model was developed to describe the accuracy of distance estimation between the lander and the landmark when the same landmark was observed twice in a row,and then the model was used to describe observability degree and to select the landmark with highest accuracy of line-of-sight depth estimation from the sequential images. The landmark selection method proposed in this paper is similar to the conventional methods in terms of navigation accuracy,but requires less optimization computation time and is more suitable for autonomous navigation on the lander.

Keywords

planetary landing / optical navigation / landmark selection / sequential images

Cite this article

Download citation ▾
LI Jiaxing, WANG Dayi, DONG Tianshu, LI Maodeng, XU Chao. Landmark Selection Method for Autonomous Optical Navigation Based on Depth Estimation Error. Journal of Deep Space Exploration, 2023, 10(2): 133‒139 https://doi.org/10.15982/j.issn.2096-9287.2023.20230019

References

[1] HUANG X Y,LI M D,WANG X L,et al. The Tianwen-1 guidance,navigation,and control for Mars entry,descent,and landing[J]. Space:Science & Technology,2021,1:1-13
[2] 徐超,王大轶,黄翔宇. 基于陆标图像的火星精确着陆自主导航方法研究[J]. 深空探测学报(中英文),2016,3(2):150-155
XU C,WANG D Y,HUANG X Y. Autonomous navigation for Mars pin-point landing based on landmark image[J]. Journal of Deep Space Exploration,2016,3(2):150-155
[3] 黄翔宇,徐超,胡荣海,等. 火星精确定点着陆多信息融合自主导航与控制方法研究[J]. 深空探测学报(中英文),2019,6(4):348-357
HUANG X Y,XU C,HU H R,et al. Research of autonomous navigation and control scheme based on multi-information fusion for mars pinpoint landing[J]. Journal of Deep Space Exploration,2019,6(4):348-357
[4] JOHNSON A,WILLSON R,CHENG Y,et al. Design through operation of an image-based velocity estimation system for Mars landing[J]. International Journal of Computer Vision,2007,74(3):319-341
[5] ULAMEC S,FANTINATI C,MAIBAUM M,et al. Rosetta lander-landing and operations on comet 67P/Churyumov-Gerasimenko[J]. Acta Astronautica,2016,125:80-91
[6] SHIRAKAWA K,MORITA H,UO M,et al. Accurate landmark tracking for navigating Hayabusa prior to final descent[C]//Proceedings of the AAS/AIAA Space Flight Mechnaics Meeting. Tampa:AIAA,2006.
[7] LU J Z,YE L L,HAN S L. Analysis and application of geometric dilution of precision based on altitude-assisted ins/sar integrated navigation[J]. IEEE Transactions on Instrumentation and Measurement,2020,70:1-10
[8] WU L,XU Q,HEIKKIL J,et al. A star sensor on-orbit calibration method based on singular value decomposition[J]. Sensors,2019,19(15):3301
[9] HUANG L,SONG J M,ZHANG C Y. Observability analysis and filter design for a vision inertial absolute navigation system for UAV using landmarks[J]. Optik,2017,149:455-468
[10] YU Z S,CUI P Y,ZHU S Y. Observability-based beacon configuration optimization for Mars entry navigation[J]. Journal of Guidance,Control,and Dynamics,2015,38(4):643-650
[11] XU C,HUANG X Y,LI M D,et al. Landmark database selection for vision-aided inertial navigation in planetary landing missions[J]. Aerospace Science and Technology,2021,118:107040
[12] ZHAO Z D,YU Z S,CUI P Y. A beacon configuration optimization method based on fisher information for Mars atmospheric entry[J]. Acta Astronautica,2017,133:467-475
[13] ZHU S Y,LIU D C,LIU Y,et al. Observability-based visual navigation using landmarks measuring angle for pinpoint landing[J]. Acta Astronautica,2019,155:313-324
[14] MOURIKIS A I,TRAWNY N,ROUMELIOTIS S I,et al. Vision-aided inertial navigation for spacecraft entry,descent,and landing[J]. IEEE Transactions on Robotics,2009,25(2):264-280
[15] HESCH J A,KOTTAS D G,BOWMAN S L,et al. Camera-IMU-based localization:observability analysis and consistency improvement[J]. The International Journal of Robotics Research,2014,33(1):182-201
[16] HESCH J A,KOTTAS D G,BOWMAN S L,et al. Consistency analysis and improvement of vision-aided inertial navigation[J]. IEEE Transactions on Robotics,2013,30(1):158-176
[17] FORSTER C,PIZZOLI M,SCARAMUZZA D. SVO:fast semi-direct monocular visual odometry[C]//Proceedings of 2014 IEEE International Conference on Robotics and Automation(ICRA). Hong Kong:ICRA,2014.
[18] PIZZOLI M,FORSTER C,SCARAMUZZA D. REMODE:probabilistic,monocular dense reconstruction in real time[C]// Proceedings of 2014 IEEE International Conference on Robotics and Automation (ICRA). Hong Kong:IEEE,2014.
[19] MARTINELLI A. Vision and IMU data fusion:closed-form solutions for attitude,speed,absolute scale,and bias determination[J]. IEEE Transactions on Robotics,2011,28(1):44-60
PDF(1342 KB)

Accesses

Citations

Detail

Sections
Recommended

/