单眼
计算机科学
人工智能
计算机视觉
适应(眼睛)
激光雷达
光流
遥感
光学
地理
图像(数学)
物理
作者
M. Shen,Z. Wang,Shuai Su,Chengju Liu,Qijun Chen
出处
期刊:IEEE Transactions on Instrumentation and Measurement
[Institute of Electrical and Electronics Engineers]
日期:2023-01-01
卷期号:72: 1-12
标识
DOI:10.1109/tim.2023.3322498
摘要
Autonomous driving necessitates ensuring safety across diverse environments, particularly in challenging conditions like low-light or nighttime scenarios. As a fundamental task in autonomous driving, monocular depth estimation has garnered significant attention and discussion. However, current monocular depth estimation methods primarily rely on daytime images, which limits their applicability to nighttime scenarios due to the substantial domain shift between daytime and nighttime styles. In this paper, we propose a novel Day-Night Adaptation method (DNA-Depth) to realize monocular depth estimation in a night environment. Specifically, we simply use Fourier Transform to address the domain alignment problem. Our method does not require extra adversarial optimization but is quite effective. The simplicity of our method makes it easy to guide day-to-night domains. To the best of our knowledge, we are the first to utilize fast Fourier transformation for nighttime monocular depth estimation. Furthermore, to alleviate the problem of mobile light sources, we utilize an unsupervised joint learning framework for depth, optical flow, and ego-motion estimation in an end-to-end manner, which are coupled by 3D geometry cues. Our model can simultaneously reason about the camera motion, the depth of a static background, and the optical flow of moving objects. Extensive experiments on the Oxford RobotCar, nuScenes, and Synthia datasets demonstrate the accuracy and precision of our method by comparing it with those state-of-the-art algorithms in depth estimation, both qualitatively and quantitatively.
科研通智能强力驱动
Strongly Powered by AbleSci AI