Representation Learning for Cross-Modality Classification

被引:9
|
作者
van Tulder, Gijs [1 ]
de Bruijne, Marleen [1 ,2 ]
机构
[1] Erasmus MC Univ, Med Ctr, Biomed Imaging Grp Rotterdam, Rotterdam, Netherlands
[2] Univ Copenhagen, Dept Comp Sci, Image Grp, Copenhagen, Denmark
关键词
Representation learning; Transfer learning; Autoencoders; Deep learning; Multi-modal image analysis; CONVOLUTIONAL NEURAL-NETWORKS; ARCHITECTURES;
D O I
10.1007/978-3-319-61188-4_12
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Differences in scanning parameters or modalities can complicate image analysis based on supervised classification. This paper presents two representation learning approaches, based on autoencoders, that address this problem by learning representations that are similar across domains. Both approaches use, next to the data representation objective, a similarity objective to minimise the difference between representations of corresponding patches from each domain. We evaluated the methods in transfer learning experiments on multi-modal brain MRI data and on synthetic data. After transforming training and test data from different modalities to the common representations learned by our methods, we trained classifiers for each of pair of modalities. We found that adding the similarity term to the standard objective can produce representations that are more similar and can give a higher accuracy in these cross-modality classification experiments.
引用
收藏
页码:126 / 136
页数:11
相关论文
共 50 条
  • [1] Representation Learning Through Cross-Modality Supervision
    Sankaran, Nishant
    Mohan, Deen Dayal
    Setlur, Srirangaraj
    Govindaraju, Venugopal
    Fedorishin, Dennis
    [J]. 2019 14TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION (FG 2019), 2019, : 107 - 114
  • [2] Cross-Modality Contrastive Learning for Hyperspectral Image Classification
    Hang, Renlong
    Qian, Xuwei
    Liu, Qingshan
    [J]. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [3] Cross-modality representation learning from transformer for hashtag prediction
    Mian Muhammad Yasir Khalil
    Qingxian Wang
    Bo Chen
    Weidong Wang
    [J]. Journal of Big Data, 10
  • [4] Cross-modality Representation Interactive Learning For Multimodal Sentiment Analysis
    Huang, Jian
    Ji, Yanli
    Yang, Yang
    Shen, Heng Tao
    [J]. PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 426 - 434
  • [5] Cross-modality representation learning from transformer for hashtag prediction
    Khalil, Mian Muhammad Yasir
    Wang, Qingxian
    Chen, Bo
    Wang, Weidong
    [J]. JOURNAL OF BIG DATA, 2023, 10 (01)
  • [6] Coral Classification Using DenseNet and Cross-modality Transfer Learning
    Xu, Lian
    Bennamoun, Mohammed
    Boussaid, Farid
    Ana, Senjian
    Sohel, Ferdous
    [J]. 2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [7] Liver Segmentation via Learning Cross-Modality Content-Aware Representation
    Lin, Xingxiao
    Ji, Zexuan
    [J]. PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT XIII, 2024, 14437 : 198 - 208
  • [8] Anatomy-Regularized Representation Learning for Cross-Modality Medical Image Segmentation
    Chen, Xu
    Lian, Chunfeng
    Wang, Li
    Deng, Hannah
    Kuang, Tianshu
    Fung, Steve
    Gateno, Jaime
    Yap, Pew-Thian
    Xia, James J.
    Shen, Dinggang
    [J]. IEEE TRANSACTIONS ON MEDICAL IMAGING, 2021, 40 (01) : 274 - 285
  • [9] Robust video question answering via contrastive cross-modality representation learning
    Xun YANG
    Jianming ZENG
    Dan GUO
    Shanshan WANG
    Jianfeng DONG
    Meng WANG
    [J]. Science China(Information Sciences)., 2024, 67 (10) - 226
  • [10] Robust video question answering via contrastive cross-modality representation learning
    Yang, Xun
    Zeng, Jianming
    Guo, Dan
    Wang, Shanshan
    Dong, Jianfeng
    Wang, Meng
    [J]. SCIENCE CHINA-INFORMATION SCIENCES, 2024, 67 (10)