Disentangling Transformer Language Models as Superposed Topic Models

被引:0
|
作者
Lim, Jia Peng [1 ]
Lauw, Hady W. [1 ]
机构
[1] Singapore Management Univ, Singapore, Singapore
基金
新加坡国家研究基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Topic Modelling is an established research area where the quality of a given topic is measured using coherence metrics. Often, we infer topics from Neural Topic Models (NTM) by interpreting their decoder weights, consisting of top-activated words projected from individual neurons. Transformer-based Language Models (TLM) similarly consist of decoder weights. However, due to its hypothesised superposition properties, the final logits originating from the residual path are considered uninterpretable. Therefore, we posit that we can interpret TLM as superposed NTM by proposing a novel weight-based, model-agnostic and corpus-agnostic approach to search and disentangle decoder-only TLM, potentially mapping individual neurons to multiple coherent topics. Our results show that it is empirically feasible to disentangle coherent topics from GPT-2 models using the Wikipedia corpus. We validate this approach for GPT-2 models using Zero-Shot Topic Modelling. Finally, we extend the proposed approach to disentangle and analyse LLaMA models.
引用
收藏
页码:8646 / 8666
页数:21
相关论文
共 50 条
  • [31] Revisiting Automated Topic Model Evaluation with Large Language Models
    Stammbach, Dominik
    Zouhar, Vilem
    Hoyle, Alexander
    Sachan, Mrinmaya
    Ash, Elliott
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 9348 - 9357
  • [32] Topic adaptation for language modeling using unnormalized exponential models
    Chen, SF
    Seymore, K
    Rosenfeld, R
    PROCEEDINGS OF THE 1998 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, VOLS 1-6, 1998, : 681 - 684
  • [33] Topic based language models for ad hoc information retrieval
    Azzopardi, L
    Girolami, M
    van Rijsbergen, CJ
    2004 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-4, PROCEEDINGS, 2004, : 3281 - 3286
  • [34] Disentangling Reasoning Capabilities from Language Models with Compositional Reasoning Transformers
    Zhong, Wanjun
    Mae, Tingting
    Wang, Jiahai
    Yin, Jian
    Zhao, Tiejun
    Lin, Chin-Yew
    Duan, Nan
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 7587 - 7600
  • [35] Transformer Grammars: Augmenting Transformer Language Models with Syntactic Inductive Biases at Scale
    Sartran, Laurent
    Barrett, Samuel
    Kuncoro, Adhiguna
    Stanojevic, Milos
    Blunsom, Phil
    Dyer, Chris
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2022, 10 : 1423 - 1439
  • [36] MODELS FOR SUPERPOSED LAMINAR FLOW FOLDING
    ODRISCOLL, ES
    NATURE, 1962, 196 (4860) : 1146 - &
  • [37] When Topic Models Disagree: Keyphrase Extraction with Multiple Topic Models
    Sterckx, Lucas
    Demeester, Thomas
    Deleu, Johannes
    Develder, Chris
    WWW'15 COMPANION: PROCEEDINGS OF THE 24TH INTERNATIONAL CONFERENCE ON WORLD WIDE WEB, 2015, : 123 - 124
  • [38] Probabilistic generative transformer language models for generative design of molecules
    Wei, Lai
    Fu, Nihang
    Song, Yuqi
    Wang, Qian
    Hu, Jianjun
    JOURNAL OF CHEMINFORMATICS, 2023, 15 (01)
  • [39] MIXED PRECISION QUANTIZATION OF TRANSFORMER LANGUAGE MODELS FOR SPEECH RECOGNITION
    Xu, Junhao
    Hu, Shoukang
    Yu, Jianwei
    Liu, Xunying
    Meng, Helen
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7383 - 7387
  • [40] Transformer Language Models Handle Word Frequency in Prediction Head
    Kobayashi, Goro
    Kuribayashi, Tatsuki
    Yokoi, Sho
    Inui, Kentaro
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 4523 - 4535