Disentangling Transformer Language Models as Superposed Topic Models

被引:0
|
作者
Lim, Jia Peng [1 ]
Lauw, Hady W. [1 ]
机构
[1] Singapore Management Univ, Singapore, Singapore
基金
新加坡国家研究基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Topic Modelling is an established research area where the quality of a given topic is measured using coherence metrics. Often, we infer topics from Neural Topic Models (NTM) by interpreting their decoder weights, consisting of top-activated words projected from individual neurons. Transformer-based Language Models (TLM) similarly consist of decoder weights. However, due to its hypothesised superposition properties, the final logits originating from the residual path are considered uninterpretable. Therefore, we posit that we can interpret TLM as superposed NTM by proposing a novel weight-based, model-agnostic and corpus-agnostic approach to search and disentangle decoder-only TLM, potentially mapping individual neurons to multiple coherent topics. Our results show that it is empirically feasible to disentangle coherent topics from GPT-2 models using the Wikipedia corpus. We validate this approach for GPT-2 models using Zero-Shot Topic Modelling. Finally, we extend the proposed approach to disentangle and analyse LLaMA models.
引用
收藏
页码:8646 / 8666
页数:21
相关论文
共 50 条
  • [21] Topic Classification for Political Texts with Pretrained Language Models
    Wang, Yu
    POLITICAL ANALYSIS, 2023, 31 (04) : 662 - 668
  • [22] Probabilistic Learning Models for Topic Extraction in Thai Language
    Asawaroengchai, Chulayuth
    Chaisangmongkon, Warasinee
    Laowattana, Djitt
    PROCEEDINGS OF 2018 5TH INTERNATIONAL CONFERENCE ON BUSINESS AND INDUSTRIAL RESEARCH (ICBIR): SMART TECHNOLOGY FOR NEXT GENERATION OF INFORMATION, ENGINEERING, BUSINESS AND SOCIAL SCIENCE, 2018, : 35 - 40
  • [23] Topic signature language models for ad hoc retrieval
    Zhou, Xiaohua
    Hu, Xiaohua
    Zhang, Xiaodan
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2007, 19 (09) : 1276 - 1287
  • [24] Beyond Topic Modeling: Comparative Evaluation of Topic Interpretation by Large Language Models
    de Melo, Tiago
    Merialdo, Paolo
    INTELLIGENT SYSTEMS AND APPLICATIONS, VOL 4, INTELLISYS 2024, 2024, 1068 : 215 - 230
  • [25] Propositional Reasoning via Neural Transformer Language Models
    Tomasic, Anthony
    Romero, Oscar J.
    Zimmerman, John
    Steinfeld, Aaron
    NEURAL-SYMBOLIC LEARNING AND REASONING, NESY 2022, 2022, : 104 - 119
  • [26] Can Transformer Language Models Predict Psychometric Properties?
    Laverghetta, Antonio, Jr.
    Nighojkar, Animesh
    Mirzakhalov, Jamshidbek
    Licato, John
    10TH CONFERENCE ON LEXICAL AND COMPUTATIONAL SEMANTICS (SEM 2021), 2021, : 12 - 25
  • [27] Improved Hybrid Streaming ASR with Transformer Language Models
    Baquero-Arnal, Pau
    Jorge, Javier
    Gimenez, Adria
    Albert Silvestre-Cerda, Joan
    Iranzo-Sanchez, Javier
    Sanchis, Albert
    Civera, Jorge
    Juan, Alfons
    INTERSPEECH 2020, 2020, : 2127 - 2131
  • [28] Disentangling frequency models
    Gomes, Erika
    Gzyl, Henryk
    JOURNAL OF OPERATIONAL RISK, 2014, 9 (02): : 3 - 21
  • [29] Topic-based language models using dirichlet mixtures
    Sadamitsu, Kugatsu
    Mishina, Takuya
    Yamamoto, Mikio
    Systems and Computers in Japan, 2007, 38 (12): : 76 - 85
  • [30] Collaborative Personalized Twitter Search with Topic-Language Models
    Vosecky, Jan
    Leung, Kenneth Wai-Ting
    Ng, Wilfred
    SIGIR'14: PROCEEDINGS OF THE 37TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2014, : 53 - 62