Joint Versus Independent Multiview Hashing for Cross-View Retrieval

被引:22
|
作者
Hu, Peng [1 ,2 ]
Peng, Xi [1 ]
Zhu, Hongyuan [2 ]
Lin, Jie [2 ]
Zhen, Liangli [3 ]
Peng, Dezhong [1 ,4 ,5 ]
机构
[1] Sichuan Univ, Coll Comp Sci, Chengdu 610065, Peoples R China
[2] Agcy Sci Technol & Res, Inst Infocomm Res, Singapore, Singapore
[3] Agcy Sci Technol & Res, Inst High Performance Comp, Singapore, Singapore
[4] Frontier Acad Ctr, Shenzhen Peng Cheng Lab, Shenzhen 518052, Peoples R China
[5] Southwest Univ, Coll Comp & Informat Sci, Chongqing 400715, Peoples R China
基金
中国国家自然科学基金;
关键词
Semantics; Decoding; Training; Computer science; Kernel; Logistics; Cybernetics; Common hamming space; cross-view retrieval; decoupled cross-view hashing network (DCHN); multiview hashing; multiview representation learning; NETWORK; SPARSE;
D O I
10.1109/TCYB.2020.3027614
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Thanks to the low storage cost and high query speed, cross-view hashing (CVH) has been successfully used for similarity search in multimedia retrieval. However, most existing CVH methods use all views to learn a common Hamming space, thus making it difficult to handle the data with increasing views or a large number of views. To overcome these difficulties, we propose a decoupled CVH network (DCHN) approach which consists of a semantic hashing autoencoder module (SHAM) and multiple multiview hashing networks (MHNs). To be specific, SHAM adopts a hashing encoder and decoder to learn a discriminative Hamming space using either a few labels or the number of classes, that is, the so-called flexible inputs. After that, MHN independently projects all samples into the discriminative Hamming space that is treated as an alternative ground truth. In brief, the Hamming space is learned from the semantic space induced from the flexible inputs, which is further used to guide view-specific hashing in an independent fashion. Thanks to such an independent/decoupled paradigm, our method could enjoy high computational efficiency and the capacity of handling the increasing number of views by only using a few labels or the number of classes. For a newly coming view, we only need to add a view-specific network into our model and avoid retraining the entire model using the new and previous views. Extensive experiments are carried out on five widely used multiview databases compared with 15 state-of-the-art approaches. The results show that the proposed independent hashing paradigm is superior to the common joint ones while enjoying high efficiency and the capacity of handling newly coming views.
引用
收藏
页码:4982 / 4993
页数:12
相关论文
共 50 条
  • [21] Multiview max-margin subspace learning for cross-view gait recognition
    Xu, Wanjiang
    Zhu, Canyan
    Wang, Ziou
    PATTERN RECOGNITION LETTERS, 2018, 107 : 75 - 82
  • [22] Cross-View Gait Recognition Using Joint Bayesian
    Li, Chao
    Sun, Shouqian
    Chen, Xiaoyu
    Min, Xin
    NINTH INTERNATIONAL CONFERENCE ON DIGITAL IMAGE PROCESSING (ICDIP 2017), 2017, 10420
  • [23] Hybrid Contrastive Quantization for Efficient Cross-View Video Retrieval
    Wang, Jinpeng
    Chen, Bin
    Liao, Dongliang
    Zeng, Ziyun
    Li, Gongfu
    Xia, Shu-Tao
    Xu, Jin
    PROCEEDINGS OF THE ACM WEB CONFERENCE 2022 (WWW'22), 2022, : 3020 - 3030
  • [24] Multiview Co-segmentation for Wide Baseline Images using Cross-view Supervision
    Yao, Yuan
    Park, Hyun Soo
    2020 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2020, : 1931 - 1940
  • [25] Cross-View Down/Up-Sampling Method for Multiview Depth Video Coding
    Liu, Qiong
    Yang, You
    Ji, Rongrong
    Gao, Yue
    Yu, Li
    IEEE SIGNAL PROCESSING LETTERS, 2012, 19 (05) : 295 - 298
  • [26] Cross-View Image Retrieval - Ground to Aerial Image Retrieval Through Deep Learning
    Khurshid, Numan
    Hanif, Talha
    Tharani, Mohbat
    Taj, Murtaza
    NEURAL INFORMATION PROCESSING (ICONIP 2019), PT II, 2019, 11954 : 210 - 221
  • [27] SimH: A Supervised Cross-View Hashing Framework Preserving Semantic Similarities in Hamming Space
    Xia Shijun
    Gu Zhongyuan
    Ge Shengbin
    Hu Weijin
    8TH INTERNATIONAL CONFERENCE ON INTERNET MULTIMEDIA COMPUTING AND SERVICE (ICIMCS2016), 2016, : 217 - 222
  • [28] Scene-Centric Joint Parsing of Cross-View Videos
    Qi, Hang
    Xu, Yuanlu
    Yuan, Tao
    Wu, Tianfu
    Zhu, Song-Chun
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 7292 - 7299
  • [29] Joint feature fusion hashing for cross-modal retrieval
    Cao, Yuxia
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (12) : 6149 - 6162
  • [30] MVCV-Traffic: multiview road traffic state estimation via cross-view learning
    Deng, Min
    Chen, Kaiqi
    Lei, Kaiyuan
    Chen, Yuanfang
    Shi, Yan
    INTERNATIONAL JOURNAL OF GEOGRAPHICAL INFORMATION SCIENCE, 2023, 37 (10) : 2205 - 2237