Disentangling Transformer Language Models as Superposed Topic Models

被引:0
|
作者
Lim, Jia Peng [1 ]
Lauw, Hady W. [1 ]
机构
[1] Singapore Management Univ, Singapore, Singapore
基金
新加坡国家研究基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Topic Modelling is an established research area where the quality of a given topic is measured using coherence metrics. Often, we infer topics from Neural Topic Models (NTM) by interpreting their decoder weights, consisting of top-activated words projected from individual neurons. Transformer-based Language Models (TLM) similarly consist of decoder weights. However, due to its hypothesised superposition properties, the final logits originating from the residual path are considered uninterpretable. Therefore, we posit that we can interpret TLM as superposed NTM by proposing a novel weight-based, model-agnostic and corpus-agnostic approach to search and disentangle decoder-only TLM, potentially mapping individual neurons to multiple coherent topics. Our results show that it is empirically feasible to disentangle coherent topics from GPT-2 models using the Wikipedia corpus. We validate this approach for GPT-2 models using Zero-Shot Topic Modelling. Finally, we extend the proposed approach to disentangle and analyse LLaMA models.
引用
收藏
页码:8646 / 8666
页数:21
相关论文
共 50 条
  • [1] Using Language Models and Topic Models for XML Retrieval
    Huang, Fang
    FOCUSED ACCESS TO XML DOCUMENTS, 2008, 4862 : 94 - 102
  • [2] LVCSR with Transformer Language Models
    Beck, Eugen
    Schlueter, Ralf
    Ney, Hermann
    INTERSPEECH 2020, 2020, : 1798 - 1802
  • [3] Empirical Study of Tweets Topic Classification Using Transformer-Based Language Models
    Mandal, Ranju
    Chen, Jinyan
    Becken, Susanne
    Stantic, Bela
    INTELLIGENT INFORMATION AND DATABASE SYSTEMS, ACIIDS 2021, 2021, 12672 : 340 - 350
  • [4] Tweets Topic Classification and Sentiment Analysis Based on Transformer-Based Language Models
    Mandal, Ranju
    Chen, Jinyan
    Becken, Susanne
    Stantic, Bela
    VIETNAM JOURNAL OF COMPUTER SCIENCE, 2023, 10 (02) : 117 - 134
  • [5] Combining Topic Specific Language Models
    Shi, Yangyang
    Wiggers, Pascal
    Jonker, Catholijn M.
    TEXT, SPEECH AND DIALOGUE, TSD 2011, 2011, 6836 : 99 - 106
  • [6] Topic Models Over Spoken Language
    Pansare, Niketan
    Jermaine, Chris
    Haas, Peter
    Rajput, Nitendra
    12TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2012), 2012, : 1062 - 1067
  • [7] When Language Models Fall in Love: Animacy Processing in Transformer Language Models
    Hanna, Michael
    Belinkov, Yonatan
    Pezzelle, Sandro
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 12120 - 12135
  • [8] Structural Guidance for Transformer Language Models
    Qian, Peng
    Naseem, Tahira
    Levy, Roger
    Astudillo, Ramon Fernandez
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 3735 - 3745
  • [9] Molecular language models: RNNs or transformer?
    Chen, Yangyang
    Wang, Zixu
    Zeng, Xiangxiang
    Li, Yayang
    Li, Pengyong
    Ye, Xiucai
    Sakurai, Tetsuya
    BRIEFINGS IN FUNCTIONAL GENOMICS, 2023, 22 (04) : 392 - 400
  • [10] Staged Training for Transformer Language Models
    Shen, Sheng
    Walsh, Pete
    Keutzer, Kurt
    Dodge, Jesse
    Peters, Matthew
    Beltagy, Iz
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,