BottleSum: Unsupervised and Self-supervised Sentence Summarization using the Information Bottleneck Principle

被引:0
|
作者
West, Peter [1 ]
Holtzman, Ari [1 ,2 ]
Buys, Jan [1 ]
Choi, Yejin [1 ,2 ]
机构
[1] Univ Washington, Paul G Allen Sch Comp Sci & Engn, Seattle, WA 98195 USA
[2] Allen Inst Artificial Intelligence, Seattle, WA USA
基金
加拿大自然科学与工程研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The principle of the Information Bottleneck (Tishby et al., 1999) is to produce a summary of information X optimized to predict some other relevant information Y. In this paper, we propose a novel approach to unsupervised sentence summarization by mapping the Information Bottleneck principle to a conditional language modelling objective: given a sentence, our approach seeks a compressed sentence that can best predict the next sentence. Our iterative algorithm under the Information Bottleneck objective searches gradually shorter subsequences of the given sentence while maximizing the probability of the next sentence conditioned on the summary. Using only pretrained language models with no direct supervision, our approach can efficiently perform extractive sentence summarization over a large corpus. Building on our unsupervised extractive summarization (BottleSumEx), we then present a new approach to self-supervised abstractive summarization (BottleSumSelf), where a transformer-based language model is trained on the output summaries of our unsupervised method. Empirical results demonstrate that our extractive method outperforms other unsupervised models on multiple automatic metrics. In addition, we find that our selfsupervised abstractive model outperforms unsupervised baselines (including our own) by human evaluation along multiple attributes.
引用
收藏
页码:3752 / 3761
页数:10
相关论文
共 50 条
  • [1] Self-Supervised Graph Representation Learning via Information Bottleneck
    Gu, Junhua
    Zheng, Zichen
    Zhou, Wenmiao
    Zhang, Yajuan
    Lu, Zhengjun
    Yang, Liang
    SYMMETRY-BASEL, 2022, 14 (04):
  • [2] Self-Supervised Multimodal Opinion Summarization
    Im, Jinbae
    Kim, Moonki
    Lee, Hoyeop
    Cho, Hyunsouk
    Chung, Sehee
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1 (ACL-IJCNLP 2021), 2021, : 388 - 403
  • [3] Self-Supervised Graph Information Bottleneck for Multiview Molecular Embedding Learning
    Li C.
    Mao K.
    Wang S.
    Yuan Y.
    Wang G.
    IEEE Transactions on Artificial Intelligence, 2024, 5 (04): : 1554 - 1562
  • [4] Unsupervised detection of acoustic events using information bottleneck principle
    Li, Yanxiong
    Wang, Qin
    Li, Xianku
    Zhang, Xue
    Zhang, Yuhan
    Chen, Aiwu
    He, Qianhua
    Huang, Qian
    DIGITAL SIGNAL PROCESSING, 2017, 63 : 123 - 134
  • [5] ADOPTING SELF-SUPERVISED LEARNING INTO UNSUPERVISED VIDEO SUMMARIZATION THROUGH RESTORATIVE SCORE.
    Abbasi, Mehryar
    Saeedi, Parvaneh
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 425 - 429
  • [6] Dynamic Bottleneck for Robust Self-Supervised Exploration
    Bai, Chenjia
    Wang, Lingxiao
    Han, Lei
    Garg, Animesh
    Hao, Jianye
    Liu, Peng
    Wang, Zhaoran
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [7] Masked self-supervised ECG representation learning via multiview information bottleneck
    Yang, Shunxiang
    Lian, Cheng
    Zeng, Zhigang
    Xu, Bingrong
    Su, Yixin
    Xue, Chenyang
    NEURAL COMPUTING & APPLICATIONS, 2024, 36 (14): : 7625 - 7637
  • [8] Self-Supervised Information Bottleneck for Deep Multi-View Subspace Clustering
    Wang, Shiye
    Li, Changsheng
    Li, Yanming
    Yuan, Ye
    Wang, Guoren
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 (1555-1567) : 1555 - 1567
  • [9] Masked self-supervised ECG representation learning via multiview information bottleneck
    Shunxiang Yang
    Cheng Lian
    Zhigang Zeng
    Bingrong Xu
    Yixin Su
    Chenyang Xue
    Neural Computing and Applications, 2024, 36 : 7625 - 7637
  • [10] Self-Supervised Learning for Contextualized Extractive Summarization
    Wang, Hong
    Wang, Xin
    Xiong, Wenhan
    Yu, Mo
    Guo, Xiaoxiao
    Chang, Shiyu
    Wang, William Yang
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 2221 - 2227