Deep neural networks have lead to a breakthrough in depth estimation from single images. Recent work shows that the quality of these estimations is rapidly increasing. It is clear that neural networks can see depth in single images. However, to the best of our knowledge, no work currently exists that analyzes what these networks have learned. In this work we take four previously published networks and investigate what depth cues they exploit. We find that all networks ignore the apparent size of known obstacles in favor of their vertical position in the image. The use of the vertical position requires the camera pose to be known; however, we find that these networks only partially recognize changes in camera pitch and roll angles. Small changes in camera pitch are shown to disturb the estimated distance towards obstacles. The use of the vertical image position allows the networks to estimate depth towards arbitrary obstacles - even those not appearing in the training set - but may depend on features that are not universally present.
Original languageEnglish
Title of host publicationThe IEEE International Conference on Computer Vision (ICCV)
Publication statusPublished - Oct 2019
EventThe IEEE International Conference on Computer Vision 2019 - Seoul, Korea, Seoul, Korea, Republic of
Duration: 27 Oct 20192 Nov 2019


ConferenceThe IEEE International Conference on Computer Vision 2019
Abbreviated titleICCV
CountryKorea, Republic of
Internet address

    Research areas

  • neural networks, monocular depth estimation, Depth perception

ID: 69222180