Probabilistic FastText for Multi-Sense Word Embeddings

被引:0
|
作者
Athiwaratkun, Ben [1 ,4 ]
Wilson, Andrew Gordon [1 ]
Anandkumar, Anima [2 ,3 ]
机构
[1] Cornell Univ, Ithaca, NY 14853 USA
[2] AWS, Seattle, WA USA
[3] CALTECH, Pasadena, CA 91125 USA
[4] Amazon, Seattle, WA USA
关键词
D O I
暂无
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
We introduce Probabilistic FastText, a new model for word embeddings that can capture multiple word senses, sub-word structure, and uncertainty information. In particular, we represent each word with a Gaussian mixture density, where the mean of a mixture component is given by the sum of n-grams. This representation allows the model to share statistical strength across sub-word structures (e.g. Latin roots), producing accurate representations of rare, misspelt, or even unseen words. Moreover, each component of the mixture can capture a different word sense. Probabilistic FastText outperforms both FASTTEXT, which has no probabilistic model, and dictionary-level probabilistic embeddings, which do not incorporate subword structures, on several word-similarity benchmarks, including English RareWord and foreign language datasets. We also achieve state-of-art performance on benchmarks that measure ability to discern different meanings. Thus, the proposed model is the first to achieve multi-sense representations while having enriched semantics on rare words.
引用
收藏
页码:1 / 11
页数:11
相关论文
共 50 条
  • [1] Multi-sense embeddings through a word sense disambiguation process
    Ruas, Terry
    Grosky, William
    Aizawa, Akiko
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2019, 136 : 288 - 303
  • [2] Improving Document Classification with Multi-Sense Embeddings
    Gupta, Vivek
    Kumar, Ankit
    Nokhiz, Pegah
    Gupta, Harshit
    Talukdar, Partha
    [J]. ECAI 2020: 24TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, 325 : 2030 - 2037
  • [3] ADDRESSING THE POLYSEMY PROBLEM IN LANGUAGE MODELING WITH ATTENTIONAL MULTI-SENSE EMBEDDINGS
    Ma, Rao
    Jin, Lesheng
    Liu, Qi
    Chen, Lu
    Yu, Kai
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 8129 - 8133
  • [4] Towards Multi-Sense Cross-Lingual Alignment of Contextual Embeddings
    Liu, Linlin
    Nguyen, Thien Hai
    Joty, Shafiq
    Bing, Lidong
    Si, Luo
    [J]. Proceedings - International Conference on Computational Linguistics, COLING, 2022, 29 (01): : 4381 - 4396
  • [5] Towards Multi-Sense Cross-Lingual Alignment of Contextual Embeddings
    DAMO Academy, Alibaba Group, China
    不详
    [J]. arXiv,
  • [6] Multi-sense Embeddings Using Synonym Sets and Hypernym Information from Wordnet
    Mudigonda, Krishna Siva Prasad
    Sharma, Poonam
    [J]. INTERNATIONAL JOURNAL OF INTERACTIVE MULTIMEDIA AND ARTIFICIAL INTELLIGENCE, 2020, 6 (04): : 68 - 79
  • [7] EDS-MEMBED: Multi-sense embeddings based on enhanced distributional semantic structures via a graph walk over word senses
    Ayetiran, Eniafe Festus
    Sojka, Petr
    Novotny, Vit
    [J]. KNOWLEDGE-BASED SYSTEMS, 2021, 219
  • [8] Adaptive GloVe and FastText Model for Hindi Word Embeddings
    Gaikwad, Vijay
    Haribhakta, Yashodhara
    [J]. PROCEEDINGS OF THE 7TH ACM IKDD CODS AND 25TH COMAD (CODS-COMAD 2020), 2020, : 175 - 179
  • [9] Extending Multi-Sense Word Embedding to Phrases and Sentences for Unsupervised Semantic Applications
    Chang, Haw-Shivan
    Agrawal, Amol
    McCallum, Andrew
    [J]. THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 6956 - 6965
  • [10] A Knowledge-Enriched Ensemble Method for Word Embedding and Multi-Sense Embedding
    Fang, Lanting
    Luo, Yong
    Feng, Kaiyu
    Zhao, Kaiqi
    Hu, Aiqun
    [J]. IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (06) : 5534 - 5549