BottleSum: Unsupervised and Self-supervised Sentence Summarization using the Information Bottleneck Principle

被引:0
|
作者
West, Peter [1 ]
Holtzman, Ari [1 ,2 ]
Buys, Jan [1 ]
Choi, Yejin [1 ,2 ]
机构
[1] Univ Washington, Paul G Allen Sch Comp Sci & Engn, Seattle, WA 98195 USA
[2] Allen Inst Artificial Intelligence, Seattle, WA USA
基金
加拿大自然科学与工程研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The principle of the Information Bottleneck (Tishby et al., 1999) is to produce a summary of information X optimized to predict some other relevant information Y. In this paper, we propose a novel approach to unsupervised sentence summarization by mapping the Information Bottleneck principle to a conditional language modelling objective: given a sentence, our approach seeks a compressed sentence that can best predict the next sentence. Our iterative algorithm under the Information Bottleneck objective searches gradually shorter subsequences of the given sentence while maximizing the probability of the next sentence conditioned on the summary. Using only pretrained language models with no direct supervision, our approach can efficiently perform extractive sentence summarization over a large corpus. Building on our unsupervised extractive summarization (BottleSumEx), we then present a new approach to self-supervised abstractive summarization (BottleSumSelf), where a transformer-based language model is trained on the output summaries of our unsupervised method. Empirical results demonstrate that our extractive method outperforms other unsupervised models on multiple automatic metrics. In addition, we find that our selfsupervised abstractive model outperforms unsupervised baselines (including our own) by human evaluation along multiple attributes.
引用
收藏
页码:3752 / 3761
页数:10
相关论文
共 50 条
  • [41] Contrastive Self-Supervised Learning as a Strong Baseline for Unsupervised Hashing
    Yang, Huei-Fang
    2022 IEEE 24TH INTERNATIONAL WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING (MMSP), 2022,
  • [42] Unsupervised Visual Anomaly Detection Using Self-Supervised Pre-Trained Transformer
    Kim, Jun-Hyung
    Kwon, Goo-Rak
    IEEE ACCESS, 2024, 12 : 127604 - 127613
  • [43] Self-Supervised Learning with an Information Maximization Criterion
    Ozsoy, Serdar
    Hamdan, Shadi
    Arik, Sercan O.
    Yuret, Deniz
    Erdogan, Alper T.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [44] Bayesian Self-Supervised Learning Using Local and Global Graph Information
    Polyzos, Konstantinos D.
    Sadeghi, Alireza
    Giannakis, Georgios B.
    2023 IEEE 9TH INTERNATIONAL WORKSHOP ON COMPUTATIONAL ADVANCES IN MULTI-SENSOR ADAPTIVE PROCESSING, CAMSAP, 2023, : 256 - 260
  • [45] Self-Supervised Learning for High-Resolution Remote Sensing Images Change Detection With Variational Information Bottleneck
    Wang, Congcong
    Du, Shouhang
    Sun, Wenbin
    Fan, Deqin
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2023, 16 : 5849 - 5866
  • [46] SentBench: Comprehensive Evaluation of Self-Supervised Sentence Representation with Benchmark Construction
    Liu, Xiaoming
    Lin, Hongyu
    Han, Xianpei
    Sun, Le
    CHINESE COMPUTATIONAL LINGUISTICS, CCL 2023, 2023, 14232 : 449 - 463
  • [47] Self-supervised Pre-training and Semi-supervised Learning for Extractive Dialog Summarization
    Zhuang, Yingying
    Song, Jiecheng
    Sadagopan, Narayanan
    Beniwal, Anurag
    COMPANION OF THE WORLD WIDE WEB CONFERENCE, WWW 2023, 2023, : 1069 - 1076
  • [48] An Efficient Self-Supervised Cross-View Training For Sentence Embedding
    Limkonchotiwat, Peerat
    Ponwitayarat, Wuttikorn
    Lowphansirikul, Lalita
    Udomcharoenchaikit, Can
    Chuangsuwanich, Ekapol
    Nutanong, Sarana
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2023, 11 : 1572 - 1587
  • [49] Combining Self-Training and Self-Supervised Learning for Unsupervised Disfluency Detection
    Wang, Shaolei
    Wang, Zhongyuan
    Che, Wanxiang
    Liu, Ting
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 1813 - 1822
  • [50] Distributed Compression using the Information Bottleneck Principle
    Steiner, Steffen
    Kuehn, Volker
    IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2021), 2021,