惯性参考系
增强现实
方向(向量空间)
单应性
摄像机切除
作者
David D. Diel,Paul A. DeBitetto,Seth Teller
出处
期刊:Workshop on Applications of Computer Vision
日期:2005-01-05
卷期号:2: 221-228
被引量:88
标识
DOI:10.1109/acvmot.2005.48
摘要
This paper describes a new method to improve inertial navigation using feature-based constraints from one or more video cameras. The proposed method lengthens the period of time during which a human or vehicle can navigate in GPS-deprived environments. Our approach integrates well with existing navigation systems, because we invoke general sensor models that represent a wide range of available hardware. The inertial model includes errors in bias, scale, and random walk. Any purely projective camera and tracking algorithm may be used, as long as the tracking output can be expressed as ray vectors extending from known locations on the sensor body. A modified linear Kalman filter performs the data fusion. Unlike traditional SLAM, our state vector contains only inertial sensor errors related to position. This choice allows uncertainty to be properly represented by a covariance matrix. We do not augment the state with feature coordinates. Instead, image data contributes stochastic epipolar constraints over a broad baseline in time and space, resulting in improved observability of the IMU error states. The constraints lead to a relative residual and associated relative covariance, defined partly by the state history. Navigation results are presented using high-quality synthetic data and real fisheye imagery.
科研通智能强力驱动
Strongly Powered by AbleSci AI