Delivering Meaningful Representation for Monocular Depth Estimation

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 58
  • Download : 0
Monocular depth estimation plays a key role in 3D scene understanding, and a number of recent papers have achieved significant improvements using deep learning based algorithms. Most papers among them proposed methods that use a pre-trained network as a deep feature extractor and then decode the obtained features to create a depth map. In this study, we focus on how to use this encoder-decoder structure to deliver meaningful representation throughout the entire network. We propose a new network architecture with our suggested modules to create a more accurate depth map by bridging the context between the encoding and decoding phase. First, we place the pyramid block at the bottleneck of the network to enlarge the view and convey rich information about the global context to the decoder. Second, we suggest a skip connection with the fuse module to aggregate the encoder and decoder feature. Finally, we validate our approach on the NYU Depth V2 and KITTI datasets. The experimental results show the efficacy of the suggested model and show performance gains over the state-of-the-art model.
Publisher
IEEE COMPUTER SOC
Issue Date
2021-01
Language
English
Citation

25th International Conference on Pattern Recognition (ICPR), pp.7790 - 7795

ISSN
1051-4651
DOI
10.1109/ICPR48806.2021.9412108
URI
http://hdl.handle.net/10203/288575
Appears in Collection
EE-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0