Multi-View Semi-Supervised Learning for Dialog Act Segmentation of Speech

被引:24
|
作者
Guz, Umit [1 ,2 ]
Cuendet, Sebastien [1 ,3 ]
Hakkani-Tuer, Dilek [1 ]
Tur, Gokhan [4 ]
机构
[1] Int Comp Sci Inst, Speech Grp, Berkeley, CA 94704 USA
[2] Isik Univ, Dept Elect Engn, Fac Engn, TR-34980 Istanbul, Turkey
[3] Optaros, CH-8037 Zurich, Switzerland
[4] SRI Int, Speech Technol & Res STAR Lab, Menlo Pk, CA 94025 USA
基金
瑞士国家科学基金会;
关键词
Boosting; co-training; prosody; self-training; semi-supervised learning; sentence segmentation;
D O I
10.1109/TASL.2009.2028371
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Sentence segmentation of speech aims at determining sentence boundaries in a stream of words as output by the speech recognizer. Typically, statistical methods are used for sentence segmentation. However, they require significant amounts of labeled data, preparation of which is time-consuming, labor-intensive, and expensive. This work investigates the application of multi-view semi-supervised learning algorithms on the sentence boundary classification problem by using lexical and prosodic information. The aim is to find an effective semi-supervised machine learning strategy when only small sets of sentence boundary-labeled data are available. We especially focus on two semi-supervised learning approaches, namely, self-training and co-training. We also compare different example selection strategies for co-training, namely, agreement and disagreement. Furthermore, we propose another method, called self-combined, which is a combination of self-training and co-training. The experimental results obtained on the ICSI Meeting (MRDA) Corpus show that both multi-view methods outperform self-training, and the best results are obtained using co-training alone. This study shows that sentence segmentation is very appropriate for multi-view learning since the data sets can be represented by two disjoint and redundantly sufficient feature sets, namely, using lexical and prosodic information. Performance of the lexical and prosodic models is improved by 26% and 11% relative, respectively, when only a small set of manually labeled examples is used. When both information sources are combined, the semi-supervised learning methods improve the baseline F-Measure of 69.8% to 74.2%.
引用
收藏
页码:320 / 329
页数:10
相关论文
共 50 条
  • [31] MULTI-VIEW AND MULTI-OBJECTIVE SEMI-SUPERVISED LEARNING FOR LARGE VOCABULARY CONTINUOUS SPEECH RECOGNITION
    Cui, Xiaodong
    Huang, Jing
    Chien, Jen-Tzung
    [J]. 2011 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2011, : 4668 - 4671
  • [32] A Semi-Supervised Multi-View Genetic Algorithm
    Lazarova, Gergana
    Koychev, Ivan
    [J]. 2014 2ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE, MODELLING AND SIMULATION, 2014, : 87 - 91
  • [33] Multi-view semi-supervised classification overview
    Jiang, Lekang
    [J]. PROCEEDINGS OF 2021 2ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND INFORMATION SYSTEMS (ICAIIS '21), 2021,
  • [34] Semi-supervised Multi-view Sentiment Analysis
    Lazarova, Gergana
    Koychev, Ivan
    [J]. COMPUTATIONAL COLLECTIVE INTELLIGENCE (ICCCI 2015), PT I, 2015, 9329 : 181 - 190
  • [35] Semi-supervised multi-view concept decomposition
    Jiang, Qi
    Zhou, Guoxu
    Zhao, Qibin
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2024, 241
  • [36] Semi-supervised Deep Multi-view Stereo
    Xu, Hongbin
    Chen, Weitao
    Liu, Yang
    Zhou, Zhipeng
    Xiao, Haihong
    Sun, Baigui
    Xie, Xuansong
    Kang, Wenxiong
    [J]. PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 4616 - 4625
  • [37] Latent Multi-view Semi-Supervised Classification
    Bo, Xiaofan
    Kang, Zhao
    Zhao, Zhitong
    Su, Yuanzhang
    Chen, Wenyu
    [J]. ASIAN CONFERENCE ON MACHINE LEARNING, VOL 101, 2019, 101 : 348 - 362
  • [38] A Multi-View Integrated Ensemble for the Background Discrimination of Semi-Supervised Semantic Segmentation
    Gwak, Hyunmin
    Jeong, Yongho
    Kim, Chanyeong
    Lee, Yonghak
    Yang, Seongmin
    Kim, Sunghwan
    [J]. APPLIED SCIENCES-BASEL, 2023, 13 (24):
  • [39] MMatch: Semi-Supervised Discriminative Representation Learning for Multi-View Classification
    Wang, Xiaoli
    Fu, Liyong
    Zhang, Yudong
    Wang, Yongli
    Li, Zechao
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (09) : 6425 - 6436
  • [40] Semi-supervised Unified Latent Factor learning with multi-view data
    Jiang, Yu
    Liu, Jing
    Li, Zechao
    Lu, Hanqing
    [J]. MACHINE VISION AND APPLICATIONS, 2014, 25 (07) : 1635 - 1645