Multi-modal Entity Alignment via Position-enhanced Multi-label Propagation

被引:0
|
作者
Tang, Wei [1 ]
Wang, Yuanyi [2 ,3 ]
机构
[1] Huawei Translat Serv Ctr, Beijing, Peoples R China
[2] Huawei Test, Dongguan, Guangdong, Peoples R China
[3] CRDU, ATE Dept, Dongguan, Guangdong, Peoples R China
关键词
multi-modal entity alignment; label propagation;
D O I
10.1145/3652583.3658085
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-modal Entity Alignment (MMEA) refers to utilizing multiple modalities such as text, images, videos, etc., to match entities from multiple knowledge graphs. Compared to single-modal entity alignment, multi-modal entity alignment can provide a more comprehensive description of entity semantics and improve matching accuracy. Currently, research efforts are directed towards the development of sophisticated deep learning models, such as graph neural networks, that can effectively capture and integrate the multi-modal features of entities for entity alignment tasks. While these models have shown promising results, they tend to focus on capturing only the local structure of entities, leading to the challenge of subgraph isomorphism. Moreover, the complexity of these models often hinders their scalability. To address these limitations, this paper proposes a non-neural, position-enhanced multi-modal entity alignment algorithm that leverages the label propagation technique to fuse and aggregate various multi-modal and position features, resulting in entity representations that are aware of long-term alignment information. Extensive experiments on various public datasets demonstrate that our proposed approach outperforms state-of-the-art algorithms in terms of both alignment accuracy and computational efficiency.
引用
收藏
页码:366 / 375
页数:10
相关论文
共 50 条
  • [41] Learning to Annotate Clothes in Everyday Photos: Multi-Modal, Multi-Label, Multi-Instance Approach
    Nogueira, Keiller
    Veloso, Adriano Alonso
    dos Santos, Jefersson A.
    2014 27TH SIBGRAPI CONFERENCE ON GRAPHICS, PATTERNS AND IMAGES (SIBGRAPI), 2014, : 327 - 334
  • [42] Multi-modal bilinear fusion with hybrid attention mechanism for multi-label skin lesion classification
    Wei, Yun
    Ji, Lin
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (24) : 65221 - 65247
  • [43] Micro-video multi-label classification method based on multi-modal feature encoding
    Jing P.
    Li Y.
    Su Y.
    Xi'an Dianzi Keji Daxue Xuebao/Journal of Xidian University, 2022, 49 (04): : 109 - 117
  • [44] MAFE: Multi-modal Alignment via Mutual Information Maximum Perspective in Multi-modal Fake News Detection
    Qin, Haimei
    Jing, Yaqi
    Duan, Yunqiang
    Jiang, Lei
    PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024, 2024, : 1515 - 1521
  • [45] Deep Multi-Modal Hashing With Semantic Enhancement for Multi-Label Micro-Video Retrieval
    Jing, Peiguang
    Sun, Haoyi
    Nie, Liqiang
    Li, Yun
    Su, Yuting
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (10) : 5080 - 5091
  • [46] TRANSFORMER-BASED MULTI-MODAL LEARNING FOR MULTI-LABEL REMOTE SENSING IMAGE CLASSIFICATION
    Hoffmann, David Sebastian
    Clasen, Kai Norman
    Demir, Begum
    IGARSS 2023 - 2023 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM, 2023, : 4891 - 4894
  • [47] Multi-label remote sensing classification with self-supervised gated multi-modal transformers
    Liu, Na
    Yuan, Ye
    Wu, Guodong
    Zhang, Sai
    Leng, Jie
    Wan, Lihong
    FRONTIERS IN COMPUTATIONAL NEUROSCIENCE, 2024, 18
  • [48] Context Recognition In-the-Wild: Unified Model for Multi-Modal Sensors and Multi-Label Classification
    Vaizman, Yonatan
    Weibel, Nadir
    Lanckriet, Gert
    Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 2017, 1 (04)
  • [49] MIC: Breast Cancer Multi-label Diagnostic Framework Based on Multi-modal Fusion Interaction
    Chen, Ziyan
    Yi, Sanli
    JOURNAL OF IMAGING INFORMATICS IN MEDICINE, 2025,
  • [50] Multi-modal and Multi-label Emotion Detection for Comics Based on Two-Stream Network
    Lin Z.
    Zeng B.
    Pan Z.
    Wen S.
    Zeng, Bi (zb9215@gdut.edu.cn), 1600, Science Press (34): : 1017 - 1027