人工智能
计算机科学
比例(比率)
图像(数学)
计算机视觉
地质学
地理
地图学
作者
David Eigen,Christian Puhrsch,Rob Fergus
出处
期刊:Cornell University - arXiv
日期:2014-01-01
被引量:2183
标识
DOI:10.48550/arxiv.1406.2283
摘要
Predicting depth is an essential component in understanding the 3D geometry of a scene. While for stereo images local correspondence suffices for estimation, finding depth relations from a single image is less straightforward, requiring integration of both global and local information from various cues. Moreover, the task is inherently ambiguous, with a large source of uncertainty coming from the overall scale. In this paper, we present a new method that addresses this task by employing two deep network stacks: one that makes a coarse global prediction based on the entire image, and another that refines this prediction locally. We also apply a scale-invariant error to help measure depth relations rather than scale. By leveraging the raw datasets as large sources of training data, our method achieves state-of-the-art results on both NYU Depth and KITTI, and matches detailed depth boundaries without the need for superpixelation.
科研通智能强力驱动
Strongly Powered by AbleSci AI