Multi-View Semi-Supervised Learning for Dialog Act Segmentation of Speech

被引:24
|
作者
Guz, Umit [1 ,2 ]
Cuendet, Sebastien [1 ,3 ]
Hakkani-Tuer, Dilek [1 ]
Tur, Gokhan [4 ]
机构
[1] Int Comp Sci Inst, Speech Grp, Berkeley, CA 94704 USA
[2] Isik Univ, Dept Elect Engn, Fac Engn, TR-34980 Istanbul, Turkey
[3] Optaros, CH-8037 Zurich, Switzerland
[4] SRI Int, Speech Technol & Res STAR Lab, Menlo Pk, CA 94025 USA
基金
瑞士国家科学基金会;
关键词
Boosting; co-training; prosody; self-training; semi-supervised learning; sentence segmentation;
D O I
10.1109/TASL.2009.2028371
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Sentence segmentation of speech aims at determining sentence boundaries in a stream of words as output by the speech recognizer. Typically, statistical methods are used for sentence segmentation. However, they require significant amounts of labeled data, preparation of which is time-consuming, labor-intensive, and expensive. This work investigates the application of multi-view semi-supervised learning algorithms on the sentence boundary classification problem by using lexical and prosodic information. The aim is to find an effective semi-supervised machine learning strategy when only small sets of sentence boundary-labeled data are available. We especially focus on two semi-supervised learning approaches, namely, self-training and co-training. We also compare different example selection strategies for co-training, namely, agreement and disagreement. Furthermore, we propose another method, called self-combined, which is a combination of self-training and co-training. The experimental results obtained on the ICSI Meeting (MRDA) Corpus show that both multi-view methods outperform self-training, and the best results are obtained using co-training alone. This study shows that sentence segmentation is very appropriate for multi-view learning since the data sets can be represented by two disjoint and redundantly sufficient feature sets, namely, using lexical and prosodic information. Performance of the lexical and prosodic models is improved by 26% and 11% relative, respectively, when only a small set of manually labeled examples is used. When both information sources are combined, the semi-supervised learning methods improve the baseline F-Measure of 69.8% to 74.2%.
引用
收藏
页码:320 / 329
页数:10
相关论文
共 50 条
  • [1] View Construction for Multi-view Semi-supervised Learning
    Sun, Shiliang
    Jin, Feng
    Tu, Wenting
    [J]. ADVANCES IN NEURAL NETWORKS - ISNN 2011, PT I, 2011, 6675 : 595 - 601
  • [2] Multi-view Learning for Semi-supervised Sentiment Classification
    Su, Yan
    Li, Shoushan
    Ju, Shengfeng
    Zhou, Guodong
    Li, Xiaojun
    [J]. 2012 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING (IALP 2012), 2012, : 13 - 16
  • [3] Multi-view semi-supervised learning for image classification
    Zhu, Songhao
    Sun, Xian
    Jin, Dongliang
    [J]. NEUROCOMPUTING, 2016, 208 : 136 - 142
  • [4] A Multi-view Regularization Method for Semi-supervised Learning
    Wang, Jiao
    Luo, Siwei
    Li, Yan
    [J]. ADVANCES IN NEURAL NETWORKS - ISNN 2010, PT 1, PROCEEDINGS, 2010, 6063 : 444 - 449
  • [5] BOOSTING OBJECTNESS: SEMI-SUPERVISED LEARNING FOR OBJECT DETECTION AND SEGMENTATION IN MULTI-VIEW IMAGES
    Wang, Huiling
    Wang, Tinghuai
    [J]. 2016 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING PROCEEDINGS, 2016, : 1796 - 1800
  • [6] MVPCL: multi-view prototype consistency learning for semi-supervised medical image segmentation
    Li, Xiafan
    Quan, Hongyan
    [J]. VISUAL COMPUTER, 2024,
  • [7] Fast Multi-View Semi-Supervised Learning With Learned Graph
    Zhang, Bin
    Qiang, Qianyao
    Wang, Fei
    Nie, Feiping
    [J]. IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2022, 34 (01) : 286 - 299
  • [8] SMGCL: Semi-supervised Multi-view Graph Contrastive Learning
    Zhou, Hui
    Gong, Maoguo
    Wang, Shanfeng
    Gao, Yuan
    Zhao, Zhongying
    [J]. KNOWLEDGE-BASED SYSTEMS, 2023, 260
  • [9] Multi-view semi-supervised learning with adaptive graph fusion
    Qiang, Qianyao
    Zhang, Bin
    Nie, Feiping
    Wang, Fei
    [J]. NEUROCOMPUTING, 2023, 557
  • [10] Flexible multi-view semi-supervised learning with unified graph
    Li, Zhongheng
    Qiang, Qianyao
    Zhang, Bin
    Wang, Fei
    Nie, Feiping
    [J]. NEURAL NETWORKS, 2021, 142 : 92 - 104