JOURNAL ARTICLE

SC-DepthV3: Robust Self-Supervised Monocular Depth Estimation for Dynamic Scenes

Libo SunJia-Wang BianHuangying ZhanWei YinIan ReidChunhua Shen

Year: 2023 Journal:   IEEE Transactions on Pattern Analysis and Machine Intelligence Vol: 46 (1)Pages: 497-508   Publisher: IEEE Computer Society

Abstract

Self-supervised monocular depth estimation has shown impressive results in static scenes. It relies on the multi-view consistency assumption for training networks, however, that is violated in dynamic object regions and occlusions. Consequently, existing methods show poor accuracy in dynamic scenes, and the estimated depth map is blurred at object boundaries because they are usually occluded in other training views. In this paper, we propose SC-DepthV3 for addressing the challenges. Specifically, we introduce an external pretrained monocular depth estimation model for generating single-image depth prior, namely pseudo-depth, based on which we propose novel losses to boost self-supervised training. As a result, our model can predict sharp and accurate depth maps, even when training from monocular videos of highly dynamic scenes. We demonstrate the significantly superior performance of our method over previous methods on six challenging datasets, and we provide detailed ablation studies for the proposed terms.

Keywords:
Monocular Artificial intelligence Computer science Computer vision Consistency (knowledge bases) Object (grammar) Image (mathematics) Depth map Pattern recognition (psychology)

Metrics

71
Cited By
12.92
FWCI (Field Weighted Citation Impact)
76
Refs
0.99
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Advanced Vision and Imaging
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Advanced Image Processing Techniques
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Image Processing Techniques and Applications
Physical Sciences →  Engineering →  Media Technology

Related Documents

© 2026 ScienceGate Book Chapters — All rights reserved.