Acr2Vec: Learning Acronym Representations in Twitter

被引:1
|
作者
Zhang, Zhifei [1 ,2 ,3 ,4 ]
Luo, Sheng [3 ]
Ma, Shuwen [1 ,2 ]
机构
[1] Tongji Univ, Res Ctr Big Data & Network Secur, Shanghai 200092, Peoples R China
[2] Tongji Univ, Ctr Educ Technol & Comp, Shanghai 200092, Peoples R China
[3] Tongji Univ, Dept Comp Sci & Technol, Shanghai 201804, Peoples R China
[4] Nanjing Univ, State Key Lab Novel Software Technol, Nanjing 210023, Jiangsu, Peoples R China
来源
ROUGH SETS | 2017年 / 10313卷
基金
中国国家自然科学基金;
关键词
Social media; Acronym; Representation learning; Word embeddings;
D O I
10.1007/978-3-319-60837-2_24
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Acronyms are common in Twitter and bring in new challenges to social media analysis. Distributed representations have achieved successful applications in natural language processing. An acronym is different from a single word and is generally defined by several words. To this end, we present Acr2Vec, an algorithmic framework for learning continuous representations for acronyms in Twitter. First, a Twitter ACRonym (TACR) dataset is automatically constructed, in which an acronym is expressed by one or more definitions. Then, three acronym embedding models have been proposed: MPDE (Max Pooling Definition Embedding), APDE (Average Pooling Definition Embedding), and PLAE (Paragraph-Like Acronym Embedding). The qualitative experimental results (i.e., similarity measure) and quantitative experimental results (i.e., acronym polarity classification) both show that MPDE and APDE are superior to PLAE.
引用
收藏
页码:280 / 288
页数:9
相关论文
共 50 条
  • [1] Topic2Vec: Learning Distributed Representations of Topics
    Niu, Liqiang
    Dai, Xinyu
    Zhang, Jianbing
    Chen, Jiajun
    PROCEEDINGS OF 2015 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING, 2015, : 193 - 196
  • [2] Learning Music Representations with wav2vec 2.0
    Ragano, Alessandro
    Benetos, Emmanouil
    Hines, Andrew
    2023 31ST IRISH CONFERENCE ON ARTIFICIAL INTELLIGENCE AND COGNITIVE SCIENCE, AICS, 2023,
  • [3] Wave2Vec: Learning Deep Representations for Biosignals
    Yuan, Ye
    Xun, Guangxu
    Suo, Qiuling
    Jia, Kebin
    Zhang, Aidong
    2017 17TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2017, : 1159 - 1164
  • [4] code2vec: Learning Distributed Representations of Code
    Alon, Uri
    Zilberstein, Meital
    Levy, Omer
    Yahav, Eran
    PROCEEDINGS OF THE ACM ON PROGRAMMING LANGUAGES-PACMPL, 2019, 3 (POPL):
  • [5] Spam2Vec: Learning Biased Embeddings for Spam Detection in Twitter
    Maity, Suman Kalyan
    Santosh, K. C.
    Mukherjee, Arjun
    COMPANION PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE 2018 (WWW 2018), 2018, : 63 - 64
  • [6] Con2Vec: Learning embedding representations for contrast sets
    Nguyen, Dang
    Luo, Wei
    Vo, Bay
    Nguyen, Loan T. T.
    Pedrycz, Witold
    KNOWLEDGE-BASED SYSTEMS, 2021, 229
  • [7] Commit2Vec: Learning Distributed Representations of Code Changes
    Cabrera Lozoya R.
    Baumann A.
    Sabetta A.
    Bezzi M.
    SN Computer Science, 2021, 2 (3)
  • [8] Tag2Vec: Learning Tag Representations in Tag Networks
    Wang, Junshan
    Lu, Zhicong
    Song, Guojia
    Fan, Yue
    Du, Lun
    Lin, Wei
    WEB CONFERENCE 2019: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW 2019), 2019, : 3314 - 3320
  • [9] Event2vec: Learning Representations of Events on Temporal Sequences
    Hong, Shenda
    Wu, Meng
    Li, Hongyan
    Wu, Zhengwu
    WEB AND BIG DATA, APWEB-WAIM 2017, PT II, 2017, 10367 : 33 - 47
  • [10] Post2Vec: Learning Distributed Representations of Stack Overflow Posts
    Xu, Bowen
    Thong Hoang
    Sharma, Abhishek
    Yang, Chengran
    Xia, Xin
    Lo, David
    IEEE TRANSACTIONS ON SOFTWARE ENGINEERING, 2022, 48 (09) : 3423 - 3441