Adaptive RGB Image Recognition by Visual-Depth Embedding

被引:14
|
作者
Cai, Ziyun [1 ]
Long, Yang [2 ]
Shao, Ling [3 ,4 ]
机构
[1] Nanjing Univ Posts & Telecommun, Coll Automat, Nanjing, Jiangsu, Peoples R China
[2] Newcastle Univ, Sch Comp, Open Lab, Newcastle Upon Tyne NE4 5TG, Tyne & Wear, England
[3] Incept Inst Artificial Intelligence, Abu Dhabi, U Arab Emirates
[4] Univ East Anglia, Sch Comp Sci, Norwich NR4 7TJ, Norfolk, England
关键词
RGB-D data; domain adaptation; visual categorization; NONNEGATIVE MATRIX FACTORIZATION; KERNEL;
D O I
10.1109/TIP.2018.2806839
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recognizing RGB images from RGB-D data is a promising application, which significantly reduces the cost while can still retain high recognition rates. However, existing methods still suffer from the domain shifting problem due to conventional surveillance cameras and depth sensors are using different mechanisms. In this paper, we aim to simultaneously solve the above two challenges: 1) how to take advantage of the additional depth information in the source domain? 2) how to reduce the data distribution mismatch between the source and target domains? We propose a novel method called adaptive visual-depth embedding (aVDE), which learns the compact shared latent space between two representations of labeled RGB and depth modalities in the source domain first. Then the shared latent space can help the transfer of the depth information to the unlabeled target dataset. At last, aVDE models two separate learning strategies for domain adaptation (feature matching and instance reweighting) in a unified optimization problem, which matches features and reweights instances jointly across the shared latent space and the projected target domain for an adaptive classifier. We test our method on five pairs of data sets for object recognition and scene classification, the results of which demonstrates the effectiveness of our proposed method.
引用
收藏
页码:2471 / 2483
页数:13
相关论文
共 50 条
  • [1] Adaptive Visual-Depth Fusion Transfer
    Cai, Ziyun
    Long, Yang
    Jing, Xiao-Yuan
    Shao, Ling
    COMPUTER VISION - ACCV 2018, PT IV, 2019, 11364 : 56 - 73
  • [2] LEARNING DEPTH FROM A SINGLE IMAGE USING VISUAL-DEPTH WORDS
    Kim, Sunok
    Choi, Sunghwan
    Sohn, Kwanghoon
    2015 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2015, : 1895 - 1899
  • [3] Visual-Depth Matching Network: Deep RGB-D Domain Adaptation With Unequal Categories
    Cai, Ziyun
    Jing, Xiao-Yuan
    Shao, Ling
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (06) : 4623 - 4635
  • [4] A Gesture Recognition Method Integrating RGB and Depth Image Features
    Wang, Kang
    Feng, Zhi-quan
    Ai, Chang-sheng
    Li, Ying-jun
    Han, Rui
    2018 INTERNATIONAL CONFERENCE ON COMPUTER, COMMUNICATION AND NETWORK TECHNOLOGY (CCNT 2018), 2018, 291 : 437 - 442
  • [5] A survey on deep neural networks for human action recognition in RGB image and depth image
    Wang, Hongyu
    ENERGY SCIENCE AND APPLIED TECHNOLOGY (ESAT 2016), 2016, : 697 - 703
  • [6] An adaptive image steganography based on depth-varying embedding
    He, Junhui
    Tang, Shaohua
    Wu, Tingting
    CISP 2008: FIRST INTERNATIONAL CONGRESS ON IMAGE AND SIGNAL PROCESSING, VOL 5, PROCEEDINGS, 2008, : 660 - 663
  • [7] Illusory visual-depth reversal can modulate sensations of contact surface
    Yuka Igarashi
    Keiko Omori
    Tetsuya Arai
    Yasunori Aizawa
    Experimental Brain Research, 2017, 235 : 3013 - 3022
  • [8] Visual Activity Recognition Based on Depth Contour Image
    Qian, Huimin
    Zhou, Jun
    Yuan, Yue
    Zhou, Xiaoyun
    2015 5TH INTERNATIONAL CONFERENCE ON INFORMATION SCIENCE AND TECHNOLOGY (ICIST), 2015, : 14 - 19
  • [9] Illusory visual-depth reversal can modulate sensations of contact surface
    Igarashi, Yuka
    Omori, Keiko
    Arai, Tetsuya
    Aizawa, Yasunori
    EXPERIMENTAL BRAIN RESEARCH, 2017, 235 (10) : 3013 - 3022
  • [10] RGB-D face recognition using LBP with suitable feature dimension of depth image
    Abebe, Hailay Berihu
    Hwang, Chih-Lyang
    IET CYBER-PHYSICAL SYSTEMS: THEORY & APPLICATIONS, 2019, 4 (03) : 189 - 197