TY - GEN
T1 - BGDNet
T2 - 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2024
AU - Chen, Jiajing
AU - Wan, Zhiqiang
AU - Narayana, Manjunath
AU - Li, Yuguang
AU - Hutchcroft, Will
AU - Velipasalar, Senem
AU - Kang, Sing Bing
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Depth estimation from single perspective image has received significant attention in the past decade, whereas the same task applied to single panoramic image remains comparatively under-explored. Most existing depth estimation models for panoramic images imitate models proposed for perspective images, which take RGB images as input and output depth directly. However, as demonstrated by our experiments, model performance drops significantly when the training and testing datasets greatly differ, since they over-fit the training data. To address this issue, we propose a novel method, referred to as the Background-guided Network (BGDNet), for more robust and accurate depth estimation from indoor panoramic images. Different from existing models, our proposed BGDNet first infers the background depth, namely from walls, floor and ceiling, via background masks, room layout and camera model. The background depth is then used to guide and improve the output foreground depth. We perform within dataset as well as cross-domain experiments on two benchmark datasets. The results show that BGDNet outperforms the state-of-the-art baselines, and is more robust to overfitting issues, with superior generalization across datasets.
AB - Depth estimation from single perspective image has received significant attention in the past decade, whereas the same task applied to single panoramic image remains comparatively under-explored. Most existing depth estimation models for panoramic images imitate models proposed for perspective images, which take RGB images as input and output depth directly. However, as demonstrated by our experiments, model performance drops significantly when the training and testing datasets greatly differ, since they over-fit the training data. To address this issue, we propose a novel method, referred to as the Background-guided Network (BGDNet), for more robust and accurate depth estimation from indoor panoramic images. Different from existing models, our proposed BGDNet first infers the background depth, namely from walls, floor and ceiling, via background masks, room layout and camera model. The background depth is then used to guide and improve the output foreground depth. We perform within dataset as well as cross-domain experiments on two benchmark datasets. The results show that BGDNet outperforms the state-of-the-art baselines, and is more robust to overfitting issues, with superior generalization across datasets.
KW - BGDNet
KW - Depth Estimation
KW - indoor panoramic images
UR - http://www.scopus.com/inward/record.url?scp=85206485704&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85206485704&partnerID=8YFLogxK
U2 - 10.1109/CVPRW63382.2024.00134
DO - 10.1109/CVPRW63382.2024.00134
M3 - Conference contribution
AN - SCOPUS:85206485704
T3 - IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops
SP - 1272
EP - 1281
BT - Proceedings - 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2024
PB - IEEE Computer Society
Y2 - 16 June 2024 through 22 June 2024
ER -