Multi-View Semi-Supervised Learning for Dialog Act Segmentation of Speech

被引:24
|
作者
Guz, Umit [1 ,2 ]
Cuendet, Sebastien [1 ,3 ]
Hakkani-Tuer, Dilek [1 ]
Tur, Gokhan [4 ]
机构
[1] Int Comp Sci Inst, Speech Grp, Berkeley, CA 94704 USA
[2] Isik Univ, Dept Elect Engn, Fac Engn, TR-34980 Istanbul, Turkey
[3] Optaros, CH-8037 Zurich, Switzerland
[4] SRI Int, Speech Technol & Res STAR Lab, Menlo Pk, CA 94025 USA
基金
瑞士国家科学基金会;
关键词
Boosting; co-training; prosody; self-training; semi-supervised learning; sentence segmentation;
D O I
10.1109/TASL.2009.2028371
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Sentence segmentation of speech aims at determining sentence boundaries in a stream of words as output by the speech recognizer. Typically, statistical methods are used for sentence segmentation. However, they require significant amounts of labeled data, preparation of which is time-consuming, labor-intensive, and expensive. This work investigates the application of multi-view semi-supervised learning algorithms on the sentence boundary classification problem by using lexical and prosodic information. The aim is to find an effective semi-supervised machine learning strategy when only small sets of sentence boundary-labeled data are available. We especially focus on two semi-supervised learning approaches, namely, self-training and co-training. We also compare different example selection strategies for co-training, namely, agreement and disagreement. Furthermore, we propose another method, called self-combined, which is a combination of self-training and co-training. The experimental results obtained on the ICSI Meeting (MRDA) Corpus show that both multi-view methods outperform self-training, and the best results are obtained using co-training alone. This study shows that sentence segmentation is very appropriate for multi-view learning since the data sets can be represented by two disjoint and redundantly sufficient feature sets, namely, using lexical and prosodic information. Performance of the lexical and prosodic models is improved by 26% and 11% relative, respectively, when only a small set of manually labeled examples is used. When both information sources are combined, the semi-supervised learning methods improve the baseline F-Measure of 69.8% to 74.2%.
引用
收藏
页码:320 / 329
页数:10
相关论文
共 50 条
  • [21] Inductive Multi-View Semi-supervised Learning with a Consensus Graph
    N. Ziraki
    A. Bosaghzadeh
    F. Dornaika
    Z. Ibrahim
    N. Barrena
    [J]. Cognitive Computation, 2023, 15 : 904 - 913
  • [22] Trusted Semi-Supervised Multi-View Classification With Contrastive Learning
    Wang, Xiaoli
    Wang, Yongli
    Wang, Yupeng
    Huang, Anqi
    Liu, Jun
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 8268 - 8278
  • [23] Semi-Supervised Multi-View Learning for Gene Network Reconstruction
    Ceci, Michelangelo
    Pio, Gianvito
    Kuzmanovski, Vladimir
    Dzeroski, Saso
    [J]. PLOS ONE, 2015, 10 (12):
  • [24] Semi-supervised one-pass multi-view learning
    Zhu, Changming
    Wang, Zhe
    Zhou, Rigui
    Wei, Lai
    Zhang, Xiafen
    Ding, Yi
    [J]. NEURAL COMPUTING & APPLICATIONS, 2019, 31 (11): : 8117 - 8134
  • [25] Dynamically Weighted Multi-View Semi-Supervised Learning for CAPTCHA
    He, Congqing
    Peng, Li
    Le, Yuquan
    He, Jiawei
    [J]. ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2019, PT II, 2019, 11440 : 343 - 354
  • [26] Multi-view Semi-supervised Learning for Web Image Annotation
    Hu, Mengqiu
    Yang, Yang
    Zhang, Hanwang
    Shen, Fumin
    Shao, Jie
    Zou, Fuhao
    [J]. MM'15: PROCEEDINGS OF THE 2015 ACM MULTIMEDIA CONFERENCE, 2015, : 947 - 950
  • [27] Semi-Supervised Learning for Multi-View Data Classification and Visualization
    Ziraki, Najmeh
    Bosaghzadeh, Alireza
    Dornaika, Fadi
    [J]. INFORMATION, 2024, 15 (07)
  • [28] Active Semi-Supervised Clustering based on Multi-View Learning
    Zhang, Xue
    Zhao, Dong-yan
    Wei, Shan
    Xiao, Wang-xin
    [J]. PROCEEDINGS OF THE 2009 WRI GLOBAL CONGRESS ON INTELLIGENT SYSTEMS, VOL III, 2009, : 495 - +
  • [29] Inductive Multi-View Semi-supervised Learning with a Consensus Graph
    Ziraki, N.
    Bosaghzadeh, A.
    Dornaika, F.
    Ibrahim, Z.
    Barrena, N.
    [J]. COGNITIVE COMPUTATION, 2023, 15 (03) : 904 - 913
  • [30] Multi-view semi-supervised learning for classification on dynamic networks
    Chen, Chuan
    Li, Yuzheng
    Qian, Hui
    Zheng, Zibin
    Hu, Yanqing
    [J]. KNOWLEDGE-BASED SYSTEMS, 2020, 195