Multi-modal Multi-relational Feature Aggregation Network for Medical Knowledge Representation Learning

被引:8
|
作者
Zhang, Yingying [1 ]
Fang, Quan [1 ]
Qian, Shengsheng [1 ]
Xu, Changsheng [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Automat, Univ Chinese Acad Sci, Beijing, Peoples R China
[2] Peng Cheng Lab, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
knowledge graph; heterogeneous graph; attention mechanism;
D O I
10.1145/3394171.3413736
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Representation learning of medical Knowledge Graph (KG) is an important task and forms the fundamental process for intelligent medical applications such as disease diagnosis and healthcare question answering. Therefore, many embedding models have been proposed to learn vector presentations for entities and relations but they ignore three important properties of medical KG: multi-modal, unbalanced and heterogeneous. Entities in the medical KG can carry unstructured multi-modal content, such as image and text. At the same time, the knowledge graph consists of multiple types of entities and relations, and each entity has various number of neighbors. In this paper, we propose a Multi-modal Multi-Relational Feature Aggregation Network (MMRFAN) for medical knowledge representation learning. To deal with the multi-modal content of the entity, we propose an adversarial feature learning model to map the textual and image information of the entity into the same vector space and learn the multi-modal common representation. To better capture the complex structure and rich semantics, we design a sampling mechanism and aggregate the neighbors with intra and inter-relation attention. We evaluate our model on three knowledge graphs, including FB15k-237, IMDb and Symptoms-in-Chinese with link prediction and node classification tasks. Experimental results show that our approach outperforms state-of-the-art method.
引用
收藏
页码:3956 / 3965
页数:10
相关论文
共 50 条
  • [31] Enhanced Topic Modeling with Multi-modal Representation Learning
    Zhang, Duoyi
    Wang, Yue
    Abul Bashar, Md
    Nayak, Richi
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2023, PT I, 2023, 13935 : 393 - 404
  • [32] Supervised Multi-modal Dictionary Learning for Clothing Representation
    Zhao, Qilu
    Wang, Jiayan
    Li, Zongmin
    PROCEEDINGS OF THE FIFTEENTH IAPR INTERNATIONAL CONFERENCE ON MACHINE VISION APPLICATIONS - MVA2017, 2017, : 51 - 54
  • [33] Deep contrastive representation learning for multi-modal clustering
    Lu, Yang
    Li, Qin
    Zhang, Xiangdong
    Gao, Quanxue
    NEUROCOMPUTING, 2024, 581
  • [34] Editorial for Special Issue on Multi-modal Representation Learning
    Fan, Deng-Ping
    Barnes, Nick
    Cheng, Ming-Ming
    Van Gool, Luc
    MACHINE INTELLIGENCE RESEARCH, 2024, 21 (04) : 615 - 616
  • [35] Multi-Relational Learning with Gaussian Processes
    Xu, Zhao
    Kersting, Kristian
    Tresp, Volker
    21ST INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI-09), PROCEEDINGS, 2009, : 1309 - 1314
  • [36] Multi-Relational Contrastive Learning for Recommendation
    Wei, Wei
    Xia, Lianghao
    Huang, Chao
    PROCEEDINGS OF THE 17TH ACM CONFERENCE ON RECOMMENDER SYSTEMS, RECSYS 2023, 2023, : 338 - 349
  • [37] Adaptive Convolution for Multi-Relational Learning
    Jiang, Xiaotian
    Wang, Quan
    Wang, Bin
    2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 978 - 987
  • [38] Multi-modal anchor adaptation learning for multi-modal summarization
    Chen, Zhongfeng
    Lu, Zhenyu
    Rong, Huan
    Zhao, Chuanjun
    Xu, Fan
    NEUROCOMPUTING, 2024, 570
  • [39] MRGAT: Multi-Relational Graph Attention Network for knowledge graph completion
    Dai, Guoquan
    Wang, Xizhao
    Zou, Xiaoying
    Liu, Chao
    Cen, Si
    NEURAL NETWORKS, 2022, 154 : 234 - 245
  • [40] Multi-Modal Knowledge Representation Learning via Webly-Supervised Relationships Mining
    Nian, Fudong
    Bao, Bing-Kun
    Li, Teng
    Xu, Changsheng
    PROCEEDINGS OF THE 2017 ACM MULTIMEDIA CONFERENCE (MM'17), 2017, : 411 - 419