Decoupled Word Embeddings using Latent Topics

被引:1
|
作者
Park, Heesoo [1 ]
Lee, Jongwuk [1 ]
机构
[1] Sungkyunkwan Univ, Seoul, South Korea
基金
新加坡国家研究基金会;
关键词
Multi-sense word embedding; contextualized word embedding; topic modeling;
D O I
10.1145/3341105.3373997
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we propose decoupled word embeddings (DWE) as a universal word representation that covers multiple senses of words. Toward this goal, our model represents each word as a combination of multiple word vectors that are associated with latent topics. Specifically, we decompose a word vector into multiple word vectors for multiple senses, according to the topic weight obtained from pre-trained topic models. Although this dynamic word representation is simple, the proposed model can leverage both local and global contexts. Through extensive experiments, including qualitative and quantitative analyses, we demonstrate that the proposed model is comparable to or better than state-ofthe-art word embedding models. The code is publicly available at https://github.com/righ120/DWE.
引用
收藏
页码:875 / 882
页数:8
相关论文
共 50 条
  • [1] Jointly Learning Word Embeddings and Latent Topics
    Shi, Bei
    Lam, Wai
    Jameel, Shoaib
    Schockaert, Steven
    Lai, Kwun Ping
    [J]. SIGIR'17: PROCEEDINGS OF THE 40TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2017, : 375 - 384
  • [2] LEWIS: Latent Embeddings for Word Images and their Semantics
    Gordo, Albert
    Almazan, Jon
    Murray, Naila
    Perronnin, Florent
    [J]. 2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, : 1242 - 1250
  • [3] A Latent Concept Topic Model for Robust Topic Inference Using Word Embeddings
    Hu, Weihua
    Tsujii, Jun'ichi
    [J]. PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2016), VOL 2, 2016, : 380 - 386
  • [4] Lifelong Learning of Topics and Domain-Specific Word Embeddings
    Qin, Xiaorui
    Lu, Yuyin
    Chen, Yufu
    Rao, Yanghui
    [J]. FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 2294 - 2309
  • [5] Further results on latent discourse models and word embeddings
    Khalife, Sammy
    Goncalves, Douglas
    Allouah, Youssef
    Liberti, Leo
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2021, 22
  • [6] Further results on latent discourse models and word embeddings
    Khalife, Sammy
    Gonçalves, Douglas
    Allouah, Youssef
    Liberti, Leo
    [J]. Journal of Machine Learning Research, 2021, 22
  • [7] Reactions to science communication: discovering social network topics using word embeddings and semantic knowledge
    de Lima, Bernardo Cerqueira
    Baracho, Renata Maria Abrantes
    Mandl, Thomas
    Porto, Patricia Baracho
    [J]. SOCIAL NETWORK ANALYSIS AND MINING, 2023, 13 (01)
  • [8] Mining coherent topics in documents using word embeddings and large-scale text data
    Yao, Liang
    Zhang, Yin
    Chen, Qinfei
    Qian, Hongze
    Wei, Baogang
    Hu, Zhifeng
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2017, 64 : 432 - 439
  • [9] Reactions to science communication: discovering social network topics using word embeddings and semantic knowledge
    Bernardo Cerqueira de Lima
    Renata Maria Abrantes Baracho
    Thomas Mandl
    Patricia Baracho Porto
    [J]. Social Network Analysis and Mining, 13
  • [10] Incorporating Latent Meanings of Morphological Compositions to Enhance Word Embeddings
    Xu, Yang
    Liu, Jiawei
    Yang, Wei
    Huang, Liusheng
    [J]. PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL), VOL 1, 2018, : 1232 - 1242