DnD: Dense Depth Estimation in Crowded Dynamic Indoor Scenes

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 98
  • Download : 0
We present a novel approach for estimating depth from a monocular camera as it moves through complex and crowded indoor environments, e.g., a department store or a metro station. Our approach predicts absolute scale depth maps over the entire scene consisting of a static background and multiple moving people, by training on dynamic scenes. Since it is difficult to collect dense depth maps from crowded indoor environments, we design our training framework without requiring groundtruth depths produced from depth sensing devices. Our network leverages RGB images and sparse depth maps generated from traditional 3D reconstruction methods to estimate dense depth maps. We use two constraints to handle depth for non-rigidly moving people without tracking their motion explicitly. We demonstrate that our approach offers consistent improvements over recent depth estimation methods on the NAVERLABS dataset, which includes complex and crowded scenes.
Publisher
IEEE
Issue Date
2021-10
Language
English
Citation

18th IEEE/CVF International Conference on Computer Vision (ICCV), pp.12777 - 12787

DOI
10.1109/ICCV48922.2021.01256
URI
http://hdl.handle.net/10203/289603
Appears in Collection
EE-Conference Papers(학술회의논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0