CLGLIAM: contrastive learning model based on global and local semantic interaction for address matching

被引:0
|
作者
Jianjun Lei
Chen Wu
Ying Wang
机构
[1] Chongqing University of Posts and Telecommunications,School of Computer Science and Technology
来源
Applied Intelligence | 2023年 / 53卷
关键词
Contrastive learning; Deep learning; Address matching; Semantic interaction;
D O I
暂无
中图分类号
学科分类号
摘要
As an essential part of geocoding, address matching has gained increasing research attention. Due to the long-distance dependency and unstructured property, existing address-matching methods hardly capture the contextual and implicit semantic information of unstructured text addresses. This paper presents a Contrastive Learning model based on Global and Local representation Interaction for Address Matching (referred to as CLGLIAM), which introduces a novel global and local interaction network to enhance the discrimination ability of the model on the hard negative address by associating the relationship between the global and local address representation explicitly. Simultaneously, to improve the generalization and transferability of the model, we utilize contrastive learning to enrich the data sample and extricate the model from task-specific knowledge. Furthermore, extensive experiments are conducted on Shenzhen and national address datasets to verify the effectiveness of our approach. Our model achieves state-of-the-art F1 scores of 99.26 and 98.50 on the two datasets, respectively. And the extended hard negative experiments further demonstrate the better performance of CLGLIAM in terms of semantic discrimination.
引用
收藏
页码:29267 / 29281
页数:14
相关论文
共 50 条
  • [41] Learning contrastive feature distribution model for interaction recognition
    Ji, Yanli
    Cheng, Hong
    Zheng, Yali
    Li, Haoxin
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2015, 33 : 340 - 349
  • [42] Global and Local Contrastive Learning for Self-Supervised Skeleton-Based Action Recognition
    Hu, Jinhua
    Hou, Yonghong
    Guo, Zihui
    Gao, Jiajun
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (11) : 10578 - 10589
  • [43] Contrastive Learning-Based Domain Adaptation for Semantic Segmentation
    Bhagwatkar, Rishika
    Kemekar, Saurabh
    Domatoti, Vinay
    Khan, Khursheed Munir
    Singh, Anamika
    2022 NATIONAL CONFERENCE ON COMMUNICATIONS (NCC), 2022, : 239 - 244
  • [44] Multimodal Recommendation Algorithm Based on Contrastive Learning and Semantic Enhancement
    Zhang, Kaihan
    Feng, Chenjiao
    Yao, Kaixuan
    Song, Peng
    Liang, Jiye
    Moshi Shibie yu Rengong Zhineng/Pattern Recognition and Artificial Intelligence, 2024, 37 (06): : 479 - 490
  • [45] GLSEC: Global and local semantic-enhanced contrastive framework for knowledge graph completion
    Ma, Ruixin
    Wang, Xiaoru
    Cao, Cunxi
    Bu, Xiya
    Wu, Hao
    Zhao, Liang
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 250
  • [46] LSPC: Exploring contrastive clustering based on local semantic information and prototype
    Chen, Jun-Fen
    Sun, Lang
    Xie, Bo-Jun
    INFORMATION SYSTEMS, 2024, 121
  • [47] Global and local combined contrastive learning for multi-view clustering
    Gu, Wenjie
    Zhu, Changming
    MULTIMEDIA SYSTEMS, 2024, 30 (05)
  • [48] Multi-granularity interaction model based on pinyins and radicals for Chinese semantic matching
    Zhao, Pengyu
    Lu, Wenpeng
    Wang, Shoujin
    Peng, Xueping
    Jian, Ping
    Wu, Hao
    Zhang, Weiyu
    WORLD WIDE WEB-INTERNET AND WEB INFORMATION SYSTEMS, 2022, 25 (04): : 1703 - 1723
  • [49] Multi-granularity interaction model based on pinyins and radicals for Chinese semantic matching
    Pengyu Zhao
    Wenpeng Lu
    Shoujin Wang
    Xueping Peng
    Ping Jian
    Hao Wu
    Weiyu Zhang
    World Wide Web, 2022, 25 : 1703 - 1723
  • [50] Embedding Global Contrastive and Local Location in Self-Supervised Learning
    Zhao, Wenyi
    Li, Chongyi
    Zhang, Weidong
    Yang, Lu
    Zhuang, Peixian
    Li, Lingqiao
    Fan, Kefeng
    Yang, Huihua
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (05) : 2275 - 2289