DeepSmooth: Efficient and Smooth Depth Completion
VOCVALC 2023
Sriram Krishna
Samsung Research
Bangalore, India
sriram.sk@samsung.com
Basavaraja Shanthappa Vandrotti
Samsung Research
Bangalore, India
b.vandrotti@samsung.com
Introduction
The Rationale for Depth Completion
Drawbacks of various depth estimation paradigms:
Depth Completion aims to overcome these issues by combining the best of both worlds - refining a sparse/noisy depth image with the semantic cues from an RGB image.
Contemporary Work
Our Contributions
Model Architecture - I
Model Architecture - II
A lightweight dual branch encoder-decoder, enhanced with temporal propagation, consisting of the following components:
Model Architecture - III
[Ref - A Closer Look at Spatiotemporal Convolutions for Action Recognition, Tran et al., CVPR 2018]
Loss Function - I
zflat = −1(Ax + By + D)/C
Where [A,B,C,D] are coefficients of the plane equation calculated by RANSAC.
Loss Function - II
LTPC = || Di - Dwarp+flat(i-1) ||
Results - I
[Ref - Enforcing temporal consistency in video depth estimation, Li et al., ICCVW 2021]
| Temporal Consistency ↑ | RMSE ↓ | MAE ↓ |
CostDCNet | 0.989 | 0.145 | 0.039 |
DM-LRN | 0.990 | 0.137 | 0.036 |
inDepth | 0.990 | 0.137 | 0.035 |
DeepSmooth (ours) | 0.992 | 0.142 | 0.043 |
Results - II
Results - III
Conclusion
Thank You