Combining Deep Generative Models and Multi-lingual Pretraining for Semi-supervised Document Classification

被引:0
|
作者
Zhu, Yi [1 ]
Shareghi, Ehsan [1 ,2 ]
Li, Yingzhen [3 ,5 ]
Reichart, Roi [4 ]
Korhonen, Anna [1 ]
机构
[1] Univ Cambridge, Language Technol Lab, Cambridge, England
[2] Monash Univ, Dept Data Sci & AI, Clayton, Vic, Australia
[3] Imperial Coll London, Dept Comp, London, England
[4] IIT, Technion, Fac Ind Engn & Management, Haifa, Israel
[5] Microsoft Res Cambridge, Cambridge, England
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Semi-supervised learning through deep generative models and multi-lingual pretraining techniques have orchestrated tremendous success across different areas of NLP. Nonetheless, their development has happened in isolation, while the combination of both could potentially be effective for tackling task-specific labelled data shortage. To bridge this gap, we combine semi-supervised deep generative models and multi-lingual pretraining to form a pipeline for document classification task. Compared to strong supervised learning baselines, our semi-supervised classification framework is highly competitive and outperforms the state-of-the-art counterparts in low-resource settings across several languages. (1)
引用
收藏
页码:894 / 908
页数:15
相关论文
共 50 条
  • [1] Combining deep generative and discriminative models for Bayesian semi-supervised learning
    Gordon, Jonathan
    Hernandez-Lobato, Jose Miguel
    [J]. PATTERN RECOGNITION, 2020, 100
  • [2] INDIAN BUFFET PROCESS DEEP GENERATIVE MODELS FOR SEMI-SUPERVISED CLASSIFICATION
    Chatzis, Sotirios P.
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 2456 - 2460
  • [3] Semi-supervised Learning with Deep Generative Models
    Kingma, Diederik P.
    Rezende, Danilo J.
    Mohamed, Shakir
    Welling, Max
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 27 (NIPS 2014), 2014, 27
  • [4] Classification of multi-lingual tweets, into multi-class model using Naive Bayes and semi-supervised learning
    Khan, Ayaz H.
    Zubair, Muhammad
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2020, 79 (43-44) : 32749 - 32767
  • [5] Variational Pretraining for Semi-supervised Text Classification
    Gururangan, Suchin
    Dang, Tam
    Card, Dallas
    Smith, Noah A.
    [J]. 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 5880 - 5894
  • [6] Classification of multi-lingual tweets, into multi-class model using Naïve Bayes and semi-supervised learning
    Ayaz H. Khan
    Muhammad Zubair
    [J]. Multimedia Tools and Applications, 2020, 79 : 32749 - 32767
  • [7] Semi-Supervised Analysis of the Electrocardiogram Using Deep Generative Models
    Rasmussen, Soren M.
    Jensen, Malte E. K.
    Meyhoff, Christian S.
    Aasvang, Eske K.
    Sorensen, Helge B. D.
    [J]. 2021 43RD ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE & BIOLOGY SOCIETY (EMBC), 2021, : 1124 - 1127
  • [8] Learning Disentangled Representations with Semi-Supervised Deep Generative Models
    Siddharth, N.
    Paige, Brooks
    van de Meent, Jan-Willem
    Desmaison, Alban
    Goodman, Noah D.
    Kohli, Pushmeet
    Wood, Frank
    Torr, Philip H. S.
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [9] Semi-supervised Sequential Generative Models
    Teng, Michael
    Le, Tuan Anh
    Scibior, Adam
    Wood, Frank
    [J]. CONFERENCE ON UNCERTAINTY IN ARTIFICIAL INTELLIGENCE (UAI 2020), 2020, 124 : 649 - 658
  • [10] Semi-supervised Multi-Label Topic Models for Document Classification and Sentence Labeling
    Soleimani, Hossein
    Miller, David J.
    [J]. CIKM'16: PROCEEDINGS OF THE 2016 ACM CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, 2016, : 105 - 114