Decoupled representation for multi-view learning

被引:0
|
作者
Sun, Shiding [1 ,2 ,3 ]
Wang, Bo [2 ]
Tian, Yingjie [3 ,4 ,5 ,6 ]
机构
[1] Univ Chinese Acad Sci, Sch Math Sci, Beijing 100049, Peoples R China
[2] Univ Int Business & Econ, Sch Informat Technol & Management, Beijing 100029, Peoples R China
[3] Univ Chinese Acad Sci, Sch Econ & Management, Beijing 100190, Peoples R China
[4] Chinese Acad Sci, Res Ctr Fictitious Econ & Data Sci, Beijing 100190, Peoples R China
[5] Chinese Acad Sci, Lab Big Data Min & Knowledge Management, Beijing 100190, Peoples R China
[6] UCAS, MOE Social Sci Lab Digital Econ Forecasts & Policy, Beijing 100190, Peoples R China
关键词
Multi-view learning; Representation learning; Information bottleneck; Contrastive learning;
D O I
10.1016/j.patcog.2024.110377
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning multi -view data is a central topic for advanced deep model applications. Existing efforts mainly focus on exploring shared information to maximize the consensus among all the views. However, after reasonably discarding superfluous task -irrelevant noise, the view -specific information is equally essential to downstream tasks. In this paper, we propose to decouple the multi -view representation learning into the shared and specific information extractions with parallel branches, and seamlessly adopt feature fusion in end -to -end models. The common feature is obtained based on the view -agnostic contrastive learning and view -discriminative training to minimize the discrepancy within the views. Simultaneously, the specific feature is learned with orthogonality constraints to minimize the view -level correlation. Besides, the semantic information in the features is reserved with supervised training. After disentangling the representations, we fuse the mutually complementary common and specific features for downstream tasks. Particularly, we provide a theoretical explanation for our method from an information bottleneck perspective. Compared with state-of-the-art multi -view models on benchmark datasets, we empirically demonstrate the advantage of our method in several downstream tasks, such as ordinary classification and few -shot learning. In addition, extensive experiments validate the robustness and transferability of our approach, when applying the learned representation on the source dataset to several target datasets.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Multi-view representation learning for multi-view action recognition
    Hao, Tong
    Wu, Dan
    Wang, Qian
    Sun, Jin-Sheng
    [J]. JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2017, 48 : 453 - 460
  • [2] Multi-view representation learning and understanding
    [J]. Multimedia Tools and Applications, 2021, 80 : 22865 - 22865
  • [3] Comprehensive Multi-view Representation Learning
    Zheng, Qinghai
    Zhu, Jihua
    Li, Zhongyu
    Tian, Zhiqiang
    Li, Chen
    [J]. INFORMATION FUSION, 2023, 89 : 198 - 209
  • [4] A Survey of Multi-View Representation Learning
    Li, Yingming
    Yang, Ming
    Zhang, Zhongfei
    [J]. IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2019, 31 (10) : 1863 - 1883
  • [5] On Deep Multi-View Representation Learning
    Wang, Weiran
    Arora, Raman
    Livescu, Karen
    Bilmes, Jeff
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 37, 2015, 37 : 1083 - 1092
  • [6] Robust Multi-View Representation Learning
    Venkatesan, Sibi
    Miller, James K.
    Dubrawski, Artur
    [J]. THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 13939 - 13940
  • [7] Multi-view representation learning and understanding
    Zhou, Tao
    Zhang, Yu
    Thung, Kim-Han
    Adeli, Ehsan
    Rekik, Islem
    Zhao, Qibin
    Zhang, Changqing
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2021, 80 (15) : 22865 - 22865
  • [8] DEEP MULTI-VIEW ROBUST REPRESENTATION LEARNING
    Jiao, Zhenyu
    Xu, Chao
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2017, : 2851 - 2855
  • [9] Multi-View Concept Learning for Data Representation
    Guan, Ziyu
    Zhang, Lijun
    Peng, Jinye
    Fan, Jianping
    [J]. IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2015, 27 (11) : 3016 - 3028
  • [10] Multi-View Multi-Instance Learning Based on Joint Sparse Representation and Multi-View Dictionary Learning
    Li, Bing
    Yuan, Chunfeng
    Xiong, Weihua
    Hu, Weiming
    Peng, Houwen
    Ding, Xinmiao
    Maybank, Steve
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (12) : 2554 - 2560