Prompt Tuning on Graph-Augmented Low-Resource Text Classification

被引:0
|
作者
Wen, Zhihao [1 ]
Fang, Yuan [1 ]
机构
[1] Singapore Management Univ, Sch Comp & Informat Syst, Singapore 188065, Singapore
关键词
Tuning; Text categorization; Task analysis; Accuracy; Paints; Oils; Ink; Text classification; graph; low-resource learning; pre-training; prompt;
D O I
10.1109/TKDE.2024.3440068
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Text classification is a fundamental problem in information retrieval with many real-world applications, such as predicting the topics of online articles and the categories of e-commerce product descriptions. However, low-resource text classification, with no or few labeled samples, presents a serious concern for supervised learning. Meanwhile, many text data are inherently grounded on a network structure, such as a hyperlink/citation network for online articles, and a user-item purchase network for e-commerce products. These graph structures capture rich semantic relationships, which can potentially augment low-resource text classification. In this paper, we propose a novel model called Graph-Grounded Pre-training and Prompting (G2P2) to address low-resource text classification in a two-pronged approach. During pre-training, we propose three graph interaction-based contrastive strategies to jointly pre-train a graph-text model; during downstream classification, we explore handcrafted discrete prompts and continuous prompt tuning for the jointly pre-trained model to achieve zero- and few-shot classification, respectively. Moreover, we explore the possibility of employing continuous prompt tuning for zero-shot inference. Specifically, we aim to generalize continuous prompts to unseen classes while leveraging a set of base classes. To this end, we extend G2P2 into G2P2 (& lowast;) , hinging on a new architecture of conditional prompt tuning. Extensive experiments on four real-world datasets demonstrate the strength of G2P2 in zero- and few-shot low-resource text classification tasks, and illustrate the advantage of G2P2 (& lowast; )in dealing with unseen classes.
引用
收藏
页码:9080 / 9095
页数:16
相关论文
共 50 条
  • [1] Prompt-based for Low-Resource Tibetan Text Classification
    An, Bo
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2023, 22 (08)
  • [2] The Power of Prompt Tuning for Low-Resource Semantic Parsing
    Schucher, Nathan
    Reddy, Siva
    de Vries, Harm
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022): (SHORT PAPERS), VOL 2, 2022, : 148 - 156
  • [3] Unifying Graph Retrieval and Prompt Tuning for Graph-Grounded Text Classification
    Dai, Le
    Yin, Yu
    Chen, Enhong
    Xiong, Hui
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 2682 - 2686
  • [4] Cross-Lingual Retrieval Augmented Prompt for Low-Resource Languages
    Nie, Ercong
    Liang, Sheng
    Schmid, Helmut
    Schuetze, Hinrich
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 8320 - 8340
  • [5] PromptEM: Prompt-tuning for Low-resource Generalized Entity Matching
    Wang, Pengfei
    Zeng, Xiaocan
    Chen, Lu
    Ye, Fan
    Mao, Yuren
    Zhu, Junhao
    Gao, Yunjun
    PROCEEDINGS OF THE VLDB ENDOWMENT, 2022, 16 (02): : 369 - 378
  • [6] Augmenting Low-Resource Text Classification with Graph-Grounded Pre-training and Prompting
    Wen, Zhihao
    Fang, Yuan
    PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 506 - 516
  • [7] RDF-to-Text Generation with Graph-augmented Structural Neural Encoders
    Gao, Hanning
    Wu, Lingfei
    Hu, Po
    Xu, Fangli
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 3030 - 3036
  • [8] Parameter-Efficient Low-Resource Dialogue State Tracking by Prompt Tuning
    Ma, Mingyu Derek
    Kao, Jiun-Yu
    Gao, Shuyang
    Gupta, Arpit
    Jin, Di
    Chung, Tagyoung
    Peng, Nanyun
    INTERSPEECH 2023, 2023, : 4653 - 4657
  • [9] Text Augmentation Using Dataset Reconstruction for Low-Resource Classification
    Rahamim, Adir
    Uziel, Guy
    Goldbraich, Esther
    Anaby-Tavor, Ateret
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 7389 - 7402
  • [10] "Low-Resource" Text Classification: A Parameter-Free Classification Method with Compressors
    Jiang, Zhiying
    Yang, Matthew Y. R.
    Tsirlin, Mikhail
    Tang, Raphael
    Dai, Yiqin
    Lin, Jimmy
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 6810 - 6828