Augmenting Low-Resource Text Classification with Graph-Grounded Pre-training and Prompting

被引:2
|
作者
Wen, Zhihao [1 ]
Fang, Yuan [1 ]
机构
[1] Singapore Management Univ, Singapore, Singapore
关键词
Text classification; graph neural networks; low-resource learning; pre-training; prompt-tuning;
D O I
10.1145/3539618.3591641
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Text classification is a fundamental problem in information retrieval with many real-world applications, such as predicting the topics of online articles and the categories of e-commerce product descriptions. However, low-resource text classification, with few or no labeled samples, poses a serious concern for supervised learning. Meanwhile, many text data are inherently grounded on a network structure, such as a hyperlink/citation network for online articles, and a user-item purchase network for e-commerce products. These graph structures capture rich semantic relationships, which can potentially augment low-resource text classification. In this paper, we propose a novel model called Graph-Grounded Pre-training and Prompting (G2P2) to address low-resource text classification in a two-pronged approach. During pre-training, we propose three graph interaction-based contrastive strategies to jointly pre-train a graph-text model; during downstream classification, we explore prompting for the jointly pre-trained model to achieve low-resource classification. Extensive experiments on four real-world datasets demonstrate the strength of G2P2 in zero- and few-shot low-resource text classification tasks.
引用
收藏
页码:506 / 516
页数:11
相关论文
共 50 条
  • [1] Unifying Graph Retrieval and Prompt Tuning for Graph-Grounded Text Classification
    Dai, Le
    Yin, Yu
    Chen, Enhong
    Xiong, Hui
    [J]. PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 2682 - 2686
  • [2] Investigating the Pre-Training Bias in Low-Resource Abstractive Summarization
    Chernyshev, Daniil
    Dobrov, Boris
    [J]. IEEE ACCESS, 2024, 12 : 47219 - 47230
  • [3] Pre-training on High-Resource Speech Recognition Improves Low-Resource Speech-to-Text Translation
    Bansal, Sameer
    Kamper, Herman
    Livescu, Karen
    Lopez, Adam
    Goldwater, Sharon
    [J]. 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 58 - 68
  • [4] Self-Supervised Audio-and-Text Pre-training with Extremely Low-Resource Parallel Data
    Kang, Yu
    Liu, Tianqiao
    Li, Hang
    Hao, Yang
    Ding, Wenbiao
    [J]. THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 10875 - 10883
  • [5] Low-Resource Named Entity Recognition via the Pre-Training Model
    Chen, Siqi
    Pei, Yijie
    Ke, Zunwang
    Silamu, Wushour
    [J]. SYMMETRY-BASEL, 2021, 13 (05):
  • [6] Multi-Stage Pre-training for Low-Resource Domain Adaptation
    Zhang, Rong
    Reddy, Revanth Gangi
    Sultan, Md Arafat
    Castelli, Vittorio
    Ferritto, Anthony
    Florian, Radu
    Kayi, Efsun Sarioglu
    Roukos, Salim
    Sil, Avirup
    Ward, Todd
    [J]. PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 5461 - 5468
  • [7] Pre-Training on Mixed Data for Low-Resource Neural Machine Translation
    Zhang, Wenbo
    Li, Xiao
    Yang, Yating
    Dong, Rui
    [J]. INFORMATION, 2021, 12 (03)
  • [8] Pre-training model for low-resource Chinese-Braille translation
    Yu, Hailong
    Su, Wei
    Liu, Lei
    Zhang, Jing
    Cai, Chuan
    Xu, Cunlu
    [J]. DISPLAYS, 2023, 79
  • [9] Text-to-Speech for Low-Resource Agglutinative Language With Morphology-Aware Language Model Pre-Training
    Liu, Rui
    Hu, Yifan
    Zuo, Haolin
    Luo, Zhaojie
    Wang, Longbiao
    Gao, Guanglai
    [J]. IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 1075 - 1087
  • [10] Low-Resource Neural Machine Translation Using XLNet Pre-training Model
    Wu, Nier
    Hou, Hongxu
    Guo, Ziyue
    Zheng, Wei
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2021, PT V, 2021, 12895 : 503 - 514