Graph structure prefix injection transformer for multi-modal entity alignment

被引:0
|
作者
Zhang, Yan [1 ,2 ,3 ,4 ,5 ]
Luo, Xiangyu [2 ]
Hu, Jing [2 ]
Zhang, Miao [1 ,3 ,4 ]
Xiao, Kui [1 ,3 ,4 ]
Li, Zhifei [1 ,2 ,3 ,4 ,5 ]
机构
[1] School of Computer Science, Hubei University, Wuhan,430062, China
[2] School of Cyber Science and Technology, Hubei University, Wuhan,430062, China
[3] Hubei Key Laboratory of Big Data Intelligent Analysis and Application, Hubei University, Wuhan,430062, China
[4] Key Laboratory of Intelligent Sensing System and Security (Hubei University), Ministry of Education, Wuhan,430062, China
[5] Hubei Provincial Engineering Research Center of Intelligent Connected Vehicle Network Security, Hubei University, Wuhan,430062, China
来源
Information Processing and Management | 2025年 / 62卷 / 03期
关键词
Contrastive Learning;
D O I
10.1016/j.ipm.2024.104048
中图分类号
学科分类号
摘要
Multi-modal entity alignment aims to integrate corresponding entities across different MMKGs. However, previous studies have not adequately considered the impact of graph structural heterogeneity on EA tasks. Different MMKGs typically exhibit variations in graph structural features, leading to distinct structural representations of the same entity relationships. Additionally, the topological structure of the graph also differs. To tackle these challenges, we introduce GSIEA, the MMEA framework that integrates structural prefix injection and modality fusion. Different from other methods that directly fuse structural data with multi-modal features to perform the alignment task, GSIEA separately processes structural data and multi-modal data such as images and attributes, incorporating a prefix injection interaction module within a multi-head attention mechanism to optimize the utilization of multi-modal information and minimize the impact of graph structural differences. Additionally, GSIEA employs a convolutional enhancement module to extract fine-grained multi-modal features and computes cross-modal weights to achieve feature fusion. We conduct experimental evaluations on two public datasets, containing 12,846 and 11,199 entity pairs, respectively, demonstrating that GSIEA outperforms baseline models, with an average improvement of 3.26% in MRR and a maximum gain of 12.5%. Furthermore, the average improvement in Hits@1 is 4.96%, with a maximum increase of 16.92%. The code of our model is stored at https://github.com/HubuKG/GSIEA. © 2024 Elsevier Ltd
引用
收藏
相关论文
共 50 条
  • [1] MMEA: Entity Alignment for Multi-modal Knowledge Graph
    Chen, Liyi
    Li, Zhi
    Wang, Yijun
    Xu, Tong
    Wang, Zhefeng
    Chen, Enhong
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT (KSEM 2020), PT I, 2020, 12274 : 134 - 147
  • [2] MultiJAF: Multi-modal joint entity alignment framework for multi-modal knowledge graph
    Cheng, Bo
    Zhu, Jia
    Guo, Meimei
    NEUROCOMPUTING, 2022, 500 : 581 - 591
  • [3] Multi-modal Graph Convolutional Network for Knowledge Graph Entity Alignment
    You, Yinghui
    Wei, Yuyang
    Zhang, Yanlong
    Chen, Wei
    Zhao, Lei
    WEB AND BIG DATA, PT I, APWEB-WAIM 2023, 2024, 14331 : 142 - 157
  • [4] MEAformer: Multi-modal Entity Alignment Transformer for Meta Modality Hybrid
    Chen, Zhuo
    Chen, Jiaoyan
    Zhang, Wen
    Guo, Lingbing
    Fang, Yin
    Huang, Yufeng
    Zhang, Yichi
    Geng, Yuxia
    Pan, Jeff Z.
    Song, Wenting
    Chen, Huajun
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 3317 - 3327
  • [5] Multi-modal entity alignment in hyperbolic space
    Guo, Hao
    Tang, Jiuyang
    Zeng, Weixin
    Zhao, Xiang
    Liu, Li
    NEUROCOMPUTING, 2021, 461 : 598 - 607
  • [6] Multi-modal Siamese Network for Entity Alignment
    Chen, Liyi
    Li, Zhi
    Xu, Tong
    Wu, Han
    Wang, Zhefeng
    Yuan, Nicholas Jing
    Chen, Enhong
    PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 118 - 126
  • [7] Enhanced Entity Interaction Modeling for Multi-Modal Entity Alignment
    Li, Jinxu
    Zhou, Qian
    Chen, Wei
    Zhao, Lei
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT II, KSEM 2023, 2023, 14118 : 214 - 227
  • [8] EDET: Entity Descriptor Encoder of Transformer for Multi-Modal Knowledge Graph in Scene Parsing
    Ma, Sai
    Wan, Weibing
    Yu, Zedong
    Zhao, Yuming
    APPLIED SCIENCES-BASEL, 2023, 13 (12):
  • [9] DFMKE: A dual fusion multi-modal knowledge graph embedding framework for entity alignment
    Zhu, Jia
    Huang, Changqin
    De Meo, Pasquale
    INFORMATION FUSION, 2023, 90 : 111 - 119
  • [10] Triplet-aware graph neural networks for factorized multi-modal knowledge graph entity alignment
    Li, Qian
    Li, Jianxin
    Wu, Jia
    Peng, Xutan
    Ji, Cheng
    Peng, Hao
    Wang, Lihong
    Yu, Philip S.
    NEURAL NETWORKS, 2024, 179