RDFC-GAN: RGB-Depth Fusion CycleGAN for Indoor Depth Completion

被引:0
|
作者
Wang H. [1 ]
Che Z. [2 ]
Yang Y. [1 ]
Wang M. [1 ]
Xu Z. [2 ]
Qiao X. [1 ]
Qi M. [3 ]
Feng F. [2 ]
Tang J. [2 ]
机构
[1] State Key Laboratory of Networking and Switching Technology, Beijing University of Posts and Telecommunications
[2] School of Computer Science, Beijing University of Posts and Telecommunications
关键词
Depth completion; Generative adversarial network; Indoor environment; RGB-depth fusion;
D O I
10.1109/TPAMI.2024.3388004
中图分类号
学科分类号
摘要
Raw depth images captured in indoor scenarios frequently exhibit extensive missing values due to the inherent limitations of the sensors and environments. For example, transparent materials frequently elude detection by depth sensors; surfaces may introduce measurement inaccuracies due to their polished textures, extended distances, and oblique incidence angles from the sensor. The presence of incomplete depth maps imposes significant challenges for subsequent vision applications, prompting the development of numerous depth completion techniques to mitigate this problem. Numerous methods excel at reconstructing dense depth maps from sparse samples, but they often falter when faced with extensive contiguous regions of missing depth values, a prevalent and critical challenge in indoor environments. To overcome these challenges, we design a novel two-branch end-to-end fusion network named RDFC-GAN, which takes a pair of RGB and incomplete depth images as input to predict a dense and completed depth map. The first branch employs an encoder-decoder structure, by adhering to the Manhattan world assumption and utilizing normal maps from RGB-D information as guidance, to regress the local dense depth values from the raw depth map. The other branch applies an RGB-depth fusion CycleGAN, adept at translating RGB imagery into detailed, textured depth maps while ensuring high fidelity through cycle consistency. We fuse the two branches via adaptive fusion modules named W-AdaIN and train the model with the help of pseudo depth maps. Comprehensive evaluations on NYU-Depth V2 and SUN RGB-D datasets show that our method significantly enhances depth completion performance particularly in realistic indoor settings. IEEE
引用
收藏
页码:1 / 14
页数:13
相关论文
共 50 条
  • [31] RGB-Depth Cross-Modal Person Re-identification
    Hafner, Frank M.
    Bhuiyan, Amran
    Kooij, Julian F. P.
    Granger, Eric
    2019 16TH IEEE INTERNATIONAL CONFERENCE ON ADVANCED VIDEO AND SIGNAL BASED SURVEILLANCE (AVSS), 2019,
  • [32] Robustifying semantic cognition of traversability across wearable RGB-depth cameras
    Yang, Kailun
    Bergasa, Luis M.
    Romera, Eduardo
    Wang, Kaiwei
    APPLIED OPTICS, 2019, 58 (12) : 3141 - 3155
  • [33] Human Action Recognition Using a Distributed RGB-Depth Camera Network
    Liu, Guoliang
    Tian, Guohui
    Li, Junwei
    Zhu, Xianglai
    Wang, Ziren
    IEEE SENSORS JOURNAL, 2018, 18 (18) : 7570 - 7576
  • [34] Multi-Scale Fusion Networks with RGB Image Features for Depth Map Completion
    Zheng, Bolun
    Xian, Chuhua
    Zhang, Dongjiu
    Jisuanji Fuzhu Sheji Yu Tuxingxue Xuebao/Journal of Computer-Aided Design and Computer Graphics, 2021, 33 (09): : 1407 - 1417
  • [35] Preliminary Result of a Laser-Beam Scattering RGB-Depth Sensor
    Kim, Jihoon
    Ryu, Soo-Hyun
    Hwang, Yongsop
    Doh, Nakju Lett
    2013 10TH INTERNATIONAL CONFERENCE ON UBIQUITOUS ROBOTS AND AMBIENT INTELLIGENCE (URAI), 2013, : 558 - +
  • [36] Calorie Counter: RGB-Depth Visual Estimation of Energy Expenditure at Home
    Tao, Lili
    Burghardt, Tilo
    Mirmehdi, Majid
    Damen, Dima
    Cooper, Ashley
    Hannuna, Sion
    Camplani, Massimo
    Paiement, Adeline
    Craddock, Ian
    COMPUTER VISION - ACCV 2016 WORKSHOPS, PT I, 2017, 10116 : 239 - 251
  • [37] RGB-Depth Feature for 3D Human Activity Recognition
    Zhao Yang
    Liu Zicheng
    Cheng Hong
    CHINA COMMUNICATIONS, 2013, 10 (07) : 93 - 103
  • [38] Deep-learning-ready RGB-depth images of seedling development
    Félix Mercier
    Geoffroy Couasnet
    Angelina El Ghaziri
    Nizar Bouhlel
    Alain Sarniguet
    Muriel Marchi
    Matthieu Barret
    David Rousseau
    Plant Methods, 21 (1)
  • [39] Unifying obstacle detection, recognition, and fusion based on millimeter wave radar and RGB-depth sensors for the visually impaired
    Long, Ningbo
    Wang, Kaiwei
    Cheng, Ruiqi
    Hu, Weijian
    Yang, Kailun
    REVIEW OF SCIENTIFIC INSTRUMENTS, 2019, 90 (04):
  • [40] Depth-Aided Robust Localization Approach for Relative Navigation using RGB-Depth Camera and LiDAR Sensor
    Song, Ha-Ryong
    Choi, Won-sub
    Kim, Hae-dong
    2014 INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND INFORMATION SCIENCES (ICCAIS 2014), 2014, : 105 - 110