MULTIMODAL REPRESENTATION LEARNING FOR BLASTOCYST ASSESSMENT

被引:4
|
作者
Wang, Youcheng [1 ]
Zheng, Zhe [1 ]
Ni, Na [1 ]
Tong, Guoqing [2 ]
Cheng, Nuo [3 ]
Li, Kai [3 ]
Yin, Ping [3 ]
Chen, Yuanyuan [3 ]
Wu, Yingna [1 ]
Xie, Guangping [1 ]
机构
[1] ShanghaiTech Univ, Sch Creat & Art, Ctr Adapt Syst Engn, Shanghai, Peoples R China
[2] Xi An Jiao Tong Univ, Dept Reprod Med, Affiliated Hosp 1, Xian, Peoples R China
[3] Shuguang Hosp, Reprod Med Ctr, Shanghai, Peoples R China
关键词
Blastocyst Assessment; Multimodal Representation Learning; Image-text Retrieval; Visual Transformer;
D O I
10.1109/ISBI53787.2023.10230468
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Blastocyst selection based on morphology grading is crucial in in vitro fertilization (IVF) treatment. Several research studies based on convolutional neural networks (CNNs) have been reported to select the most viable blastocyst automatically. In this paper, we propose a multimodal representation learning framework in which the text description is firstly streamed as a complementary supervision signal to enrich the visual information. Moreover, we redefine the blastocyst assessment problem to an image-text retrieval task to solve the data imbalance. The experimental results show that the performance metrics, e.g., accuracy, outperform the unimodal classification (+1.5%) and image retrieval counterparts (+1.2%), which demonstrates our proposed model's effectiveness.
引用
收藏
页数:5
相关论文
共 50 条
  • [31] Triple disentangled representation learning for multimodal affective analysis
    Zhou, Ying
    Liang, Xuefeng
    Chen, Han
    Zhao, Yin
    Chen, Xin
    Yu, Lida
    INFORMATION FUSION, 2024, 114
  • [32] Multimodal Intelligence: Representation Learning, Information Fusion, and Applications
    Zhang, Chao
    Yang, Zichao
    He, Xiaodong
    Deng, Li
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2020, 14 (03) : 478 - 493
  • [33] Multimodal Data Enhanced Representation Learning for Knowledge Graphs
    Wang, Zikang
    Li, Linjing
    Li, Qiudan
    Zeng, Daniel
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [34] Multimodal Coordinated Representation Learning Based on Evidence Theory
    Li, Wei
    Han, Deqiang
    Dezert, Jean
    Yang, Yi
    2024 27TH INTERNATIONAL CONFERENCE ON INFORMATION FUSION, FUSION 2024, 2024,
  • [35] Learning Joint Multimodal Representation with Adversarial Attention Networks
    Huang, Feiran
    Zhang, Xiaoming
    Li, Zhoujun
    PROCEEDINGS OF THE 2018 ACM MULTIMEDIA CONFERENCE (MM'18), 2018, : 1874 - 1882
  • [36] Multitask Representation Learning for Multimodal Estimation of Depression Level
    Qureshi, Syed Arbaaz
    Saha, Sriparna
    Hasanuzzaman, Mohammed
    Dias, Gael
    IEEE INTELLIGENT SYSTEMS, 2019, 34 (05) : 45 - 52
  • [37] Multimodal Fusion Representation Learning Based on Differential Privacy
    Cai, Chaoxin
    Sang, Yingpeng
    Huang, Jinghao
    Zhang, Maliang
    Li, Weizheng
    PARALLEL AND DISTRIBUTED COMPUTING, APPLICATIONS AND TECHNOLOGIES, PDCAT 2021, 2022, 13148 : 548 - 559
  • [38] Robust Representation Learning for Multimodal Emotion Recognition with Contrastive Learning and Mixup
    Car, Yunrui
    Ye, Runchuan
    Xie, Jingran
    Zhou, Yixuan
    Xu, Yaoxun
    Wu, Zhiyong
    PROCEEDINGS OF THE 2ND INTERNATIONAL WORKSHOP ON MULTIMODAL AND RESPONSIBLE AFFECTIVE COMPUTING, MRAC 2024, 2024, : 93 - 97
  • [39] Multimodal Cardiac Segmentation Using Disentangled Representation Learning
    Chartsias, Agisilaos
    Papanastasiou, Giorgos
    Wang, Chengjia
    Stirrat, Colin
    Semple, Scott
    Newby, David
    Dharmakumar, Rohan
    Tsaftaris, Sotirios A.
    STATISTICAL ATLASES AND COMPUTATIONAL MODELS OF THE HEART: MULTI-SEQUENCE CMR SEGMENTATION, CRT-EPIGGY AND LV FULL QUANTIFICATION CHALLENGES, 2020, 12009 : 128 - 137
  • [40] Deep learning with multimodal representation for pancancer prognosis prediction
    Cheerla, Anika
    Gevaert, Olivier
    BIOINFORMATICS, 2019, 35 (14) : I446 - I454