Single image depth estimation from predicted semantic labels

Beyang Liu*, Stephen Gould, Daphne Koller

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

434 Citations (Scopus)

Abstract

We consider the problem of estimating the depth of each pixel in a scene from a single monocular image. Unlike traditional approaches [18, 19], which attempt to map from appearance features to depth directly, we first perform a semantic segmentation of the scene and use the semantic labels to guide the 3D reconstruction. This approach provides several advantages: By knowing the semantic class of a pixel or region, depth and geometry constraints can be easily enforced (e.g., "sky" is far away and "ground" is horizontal). In addition, depth can be more readily predicted by measuring the difference in appearance with respect to a given semantic class. For example, a tree will have more uniform appearance in the distance than it does close up. Finally, the incorporation of semantic features allows us to achieve state-of-the-art results with a significantly simpler model than previous works.

Original languageEnglish
Title of host publication2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2010
Pages1253-1260
Number of pages8
DOIs
Publication statusPublished - 2010
Externally publishedYes
Event2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2010 - San Francisco, CA, United States
Duration: 13 Jun 201018 Jun 2010

Publication series

NameProceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition
ISSN (Print)1063-6919

Conference

Conference2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2010
Country/TerritoryUnited States
CitySan Francisco, CA
Period13/06/1018/06/10

Fingerprint

Dive into the research topics of 'Single image depth estimation from predicted semantic labels'. Together they form a unique fingerprint.

Cite this