Commonsense Knowledge Mining from Pretrained Models

被引:0
|
作者
Feldman, Joshua [1 ]
Davison, Joe [1 ]
Rush, Alexander M. [1 ]
机构
[1] Harvard Univ, Sch Engn & Appl Sci, Cambridge, MA 02138 USA
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Inferring commonsense knowledge is a key challenge in natural language processing, but due to the sparsity of training data, previous work has shown that supervised methods for commonsense knowledge mining underperform when evaluated on novel data. In this work, we develop a method for generating commonsense knowledge using a large, pre-trained bidirectional language model. By transforming relational triples into masked sentences, we can use this model to rank a triple's validity by the estimated pointwise mutual information between the two entities. Since we do not update the weights of the bidirectional model, our approach is not biased by the coverage of any one commonsense knowledge base. Though this method performs worse on a test set than models explicitly trained on a corresponding training set, it outperforms these methods when mining commonsense knowledge from new sources, suggesting that unsupervised techniques may generalize better than current supervised approaches.
引用
收藏
页码:1173 / 1178
页数:6
相关论文
共 50 条
  • [1] Enhancing pretrained language models with structured commonsense knowledge for textual inference
    Du, Li
    Ding, Xiao
    Xiong, Kai
    Liu, Ting
    Qin, Bing
    [J]. KNOWLEDGE-BASED SYSTEMS, 2022, 254
  • [2] Commonsense Knowledge Mining from the Web
    Yu, Chi-Hsin
    Chen, Hsin-Hsi
    [J]. PROCEEDINGS OF THE TWENTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE (AAAI-10), 2010, : 1480 - 1485
  • [3] Visual Commonsense in Pretrained Unimodal and Multimodal Models
    Zhang, Chenyu
    Van Durme, Benjamin
    Li, Zhuowan
    Stengel-Eskin, Elias
    [J]. NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 5321 - 5335
  • [4] Symbolic Knowledge Distillation: from General Language Models to Commonsense Models
    West, Peter
    Bhagavatula, Chandra
    Hessel, Jack
    Hwang, Jena D.
    Jiang, Liwei
    Le Bras, Ronan
    Lu, Ximing
    Welleck, Sean
    Choi, Yejin
    [J]. NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 4602 - 4625
  • [5] Mining Verb-Oriented Commonsense Knowledge
    Liu, Jingping
    Zhou, Yuanfu
    Wu, Dan
    Wang, Chao
    Jiang, Haiyun
    Zhang, Sheng
    Xu, Bo
    Xiao, Yanghua
    [J]. 2020 IEEE 36TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING (ICDE 2020), 2020, : 1830 - 1833
  • [6] Ensemble pretrained language models to extract biomedical knowledge from literature
    Li, Zhao
    Wei, Qiang
    Huang, Liang-Chin
    Li, Jianfu
    Hu, Yan
    Chuang, Yao-Shun
    He, Jianping
    Das, Avisha
    Keloth, Vipina Kuttichi
    Yang, Yuntao
    Diala, Chiamaka S.
    Roberts, Kirk E.
    Tao, Cui
    Jiang, Xiaoqian
    Zheng, W. Jim
    Xu, Hua
    [J]. JOURNAL OF THE AMERICAN MEDICAL INFORMATICS ASSOCIATION, 2024, 31 (09) : 1904 - 1911
  • [7] Eliciting Knowledge from Pretrained Language Models for Prototypical Prompt Verbalizer
    Wei, Yinyi
    Mo, Tong
    Jiang, Yongtao
    Li, Weiping
    Zhao, Wen
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2022, PT II, 2022, 13530 : 222 - 233
  • [8] GO BEYOND PLAIN FINE-TUNING: IMPROVING PRETRAINED MODELS FOR SOCIAL COMMONSENSE
    Chang, Ting-Yun
    Liu, Yang
    Gopalakrishnan, Karthik
    Hedayatnia, Behnam
    Zhou, Pei
    Hakkani-Tur, Dilek
    [J]. 2021 IEEE SPOKEN LANGUAGE TECHNOLOGY WORKSHOP (SLT), 2021, : 1028 - 1035
  • [9] Semantically Enhanced Models for Commonsense Knowledge Acquisition
    Alhussien, Ikhlas
    Cambria, Erik
    Zhang NengSheng
    [J]. 2018 18TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS (ICDMW), 2018, : 1014 - 1021
  • [10] Learning Commonsense Knowledge Models for Semantic Analytics
    Hu Shangfeng
    Kanagasabai, Rajaraman
    [J]. 2016 IEEE TENTH INTERNATIONAL CONFERENCE ON SEMANTIC COMPUTING (ICSC), 2016, : 399 - 402