Multi-Source Domain Adaptation for Text-Independent Forensic Speaker Recognition

被引:16
|
作者
Wang, Zhenyu [1 ]
Hansen, John H. L. [1 ]
机构
[1] Univ Texas Dallas, Ctr Robust Speech Syst, Erik Jonsson Sch Engn, Richardson, TX 75080 USA
关键词
Speaker recognition; Forensics; Training; Adaptation models; Acoustics; Task analysis; Speech recognition; Discrepancy loss; forensics; multi-source domain adaptation; domain adversarial training; maximum mean discrepancy; moment-matching; speaker recognition; IDENTIFICATION; FRAMEWORK;
D O I
10.1109/TASLP.2021.3130975
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Adapting speaker recognition systems to new environments is a widely-used technique to improve a well-performing model learned from large-scale data towards a task-specific small-scale data scenarios. However, previous studies focus on single domain adaptation, which neglects a more practical scenario where training data are collected from multiple acoustic domains needed in forensic scenarios. Audio analysis for forensic speaker recognition offers unique challenges in model training with multi-domain training data due to location/scenario uncertainty and diversity mismatch between reference and naturalistic field recordings. It is also difficult to directly employ small-scale domain-specific data to train complex neural network architectures due to domain mismatch and performance loss. Fine-tuning is a commonly-used method for adaptation in order to retrain the model with weights initialized from a well-trained model. Alternatively, in this study, three novel adaptation methods based on domain adversarial training, discrepancy minimization, and moment-matching approaches are proposed to further promote adaptation performance across multiple acoustic domains. A comprehensive set of experiments are conducted to demonstrate that: 1) diverse acoustic environments do impact speaker recognition performance, which could advance research in audio forensics, 2) domain adversarial training learns the discriminative features which are also invariant to shifts between domains, 3) discrepancy-minimizing adaptation achieves effective performance simultaneously across multiple acoustic domains, and 4) moment-matching adaptation along with dynamic distribution alignment also significantly promotes speaker recognition performance on each domain, especially for the LENA-field domain with noise compared to all other systems. Advancements shown here in adaptation therefore helper ensure more consistent performance for field operational data in audio forensics.
引用
收藏
页码:60 / 75
页数:16
相关论文
共 50 条
  • [41] Spin-Image Descriptors for Text-Independent Speaker Recognition
    Mohammed, Suhaila N.
    Jabir, Adnan J.
    Abbas, Zaid Ali
    EMERGING TRENDS IN INTELLIGENT COMPUTING AND INFORMATICS: DATA SCIENCE, INTELLIGENT INFORMATION SYSTEMS AND SMART COMPUTING, 2020, 1073 : 216 - 226
  • [42] FREQUENCY AND TEMPORAL CONVOLUTIONAL ATTENTION FOR TEXT-INDEPENDENT SPEAKER RECOGNITION
    Yadav, Sarthak
    Rai, Atul
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 6794 - 6798
  • [43] A Multiscale Feature Extraction Method for Text-independent Speaker Recognition
    Chen Zhigao
    Li Peng
    Xiao Runqiu
    Li Ta
    Wang Wenchao
    JOURNAL OF ELECTRONICS & INFORMATION TECHNOLOGY, 2021, 43 (11) : 3266 - 3271
  • [44] Ensemble of Support Vector Machine for Text-Independent Speaker Recognition
    Lei, Zhenchun
    Yang, Yingchun
    Wu, Zhaohui
    INTERNATIONAL JOURNAL OF COMPUTER SCIENCE AND NETWORK SECURITY, 2006, 6 (5A): : 163 - 167
  • [45] Data-model relationship in text-independent speaker recognition
    Mason, JSD
    Evans, NWD
    Stapert, R
    Auckenthaler, R
    EURASIP JOURNAL ON APPLIED SIGNAL PROCESSING, 2005, 2005 (04) : 471 - 481
  • [46] Adaptive Convolutional Neural Network for Text-Independent Speaker Recognition
    Kim, Seong-Hu
    Park, Yong-Hwa
    INTERSPEECH 2021, 2021, : 66 - 70
  • [47] Learning Vector Quantization in text-independent Automatic Speaker Recognition
    Filgueiras, TE
    Messina, RO
    Cabral, EF
    VTH BRAZILIAN SYMPOSIUM ON NEURAL NETWORKS, PROCEEDINGS, 1998, : 135 - 139
  • [48] Angular Margin Centroid Loss for Text-independent Speaker Recognition
    Wei, Yuheng
    Du, Junzhao
    Liu, Hui
    INTERSPEECH 2020, 2020, : 3820 - 3824
  • [49] An overview of text-independent speaker recognition: From features to supervectors
    Kinnunen, Tomi
    Li, Haizhou
    SPEECH COMMUNICATION, 2010, 52 (01) : 12 - 40
  • [50] Formant Trajectories in Linguistic Units for Text-Independent Speaker Recognition
    Franco-Pedroso, Javier
    Espinoza-Cuadros, Fernando
    Gonzalez-Rodriguez, Joaquin
    2013 INTERNATIONAL CONFERENCE ON BIOMETRICS (ICB), 2013,