Structure-Aware Residual Pyramid Network for Monocular Depth Estimation
Structure-Aware Residual Pyramid Network for Monocular Depth Estimation
Xiaotian Chen, Xuejin Chen, Zheng-Jun Zha
Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence
Main track. Pages 694-700.
https://doi.org/10.24963/ijcai.2019/98
Monocular depth estimation is an essential task for scene understanding. The underlying structure of objects and stuff in a complex scene is critical to recovering accurate and visually-pleasing depth maps. Global structure conveys scene layouts, while local structure reflects shape details. Recently developed approaches based on convolutional neural networks (CNNs) significantly improve the performance of depth estimation. However, few of them take into account multi-scale structures in complex scenes. In this paper, we propose a Structure-Aware Residual Pyramid Network (SARPN) to exploit multi-scale structures for accurate depth prediction. We propose a Residual Pyramid Decoder (RPD) which expresses global scene structure in upper levels to represent layouts, and local structure in lower levels to present shape details. At each level, we propose Residual Refinement Modules (RRM) that predict residual maps to progressively add finer structures on the coarser structure predicted at the upper level. In order to fully exploit multi-scale image features, an Adaptive Dense Feature Fusion (ADFF) module, which adaptively fuses effective features from all scales for inferring structures of each scale, is introduced. Experiment results on the challenging NYU-Depth v2 dataset demonstrate that our proposed approach achieves state-of-the-art performance in both qualitative and quantitative evaluation. The code is available at https://github.com/Xt-Chen/SARPN.
Keywords:
Computer Vision: 2D and 3D Computer Vision
Computer Vision: Computational Photography, Photometry, Shape from X