Publications

Detailed Information

Stad: Stable Video Depth Estimation

Cited 1 time in Web of Science Cited 1 time in Scopus
Authors

Lee, Hyunmin; Park, Jaesik

Issue Date
2021
Publisher
ICIP
Citation
Proceedings - International Conference on Image Processing, ICIP, Vol.2021-September, pp.3213-3217
Abstract
We present a method for estimating temporally stable depth video from a sequence of images. We extend the prior work aimed at video depth estimation, Neural-RGBD [1], which proposed to use temporal information by accumulating a depth probability volume over time. We propose three simple yet effective ideas to gain improvement: (1) temporal attention module to select and propagate only the meaningful temporal information, (2) geometric warping operation to warp neighbor features in the way of preserving geometry cues, and (3) scale-invariant loss to relieve the inherent scale ambiguity problem in monocular depth estimation task. We demonstrate the efficiency of proposed ideas by comparing our proposed network STAD with the state-of-the-arts. Moreover, we compare STAD with its per-frame network STAD-frame to show the importance of utilizing temporal information. The experimental results show that STAD significantly improved the baseline accuracy without a large parameter increase.
ISSN
1522-4880
URI
https://hdl.handle.net/10371/201302
DOI
https://doi.org/10.1109/ICIP42928.2021.9506521
Files in This Item:
There are no files associated with this item.
Appears in Collections:

Related Researcher

  • College of Engineering
  • Dept. of Computer Science and Engineering
Research Area Computer Graphics, Computer Vision, Machine Learning

Altmetrics

Item View & Download Count

  • mendeley

Items in S-Space are protected by copyright, with all rights reserved, unless otherwise indicated.

Share