IMPROVED LARGE-MARGIN SOFTMAX LOSS FOR SPEAKER DIARISATION

被引:0
|
作者
Fathullah, Y. [1 ]
Zhang, C. [1 ]
Woodland, P. C. [1 ]
机构
[1] Univ Cambridge, Engn Dept, Cambridge, England
关键词
Speaker diarisation; speaker embeddings; large-margin softmax; overlapping speech; DIARIZATION;
D O I
10.1109/icassp40776.2020.9053373
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Speaker diarisation systems nowadays use embeddings generated from speech segments in a bottleneck layer, which are needed to be discriminative for unseen speakers. It is well-known that large-margin training can improve the generalisation ability to unseen data, and its use in such open-set problems has been widespread. Therefore, this paper introduces a general approach to the large-margin softmax loss without any approximations to improve the quality of speaker embeddings for diarisation. Furthermore, a novel and simple way to stabilise training, when large-margin softmax is used, is proposed. Finally, to combat the effect of overlapping speech, different training margins are used to reduce the negative effect overlapping speech has on creating discriminative embeddings. Experiments on the AMI meeting corpus show that the use of large-margin softmax significantly improves the speaker error rate (SER). By using all hyper parameters of the loss in a unified way, further improvements were achieved which reached a relative SER reduction of 24.6% over the baseline. However, by training overlapping and single speaker speech samples with different margins, the best result was achieved, giving overall a 29.5% SER reduction relative to the baseline.
引用
下载
收藏
页码:7104 / 7108
页数:5
相关论文
共 50 条
  • [31] Large-margin Weakly Supervised Dimensionality Reduction
    Xu, Chang
    Tao, Dacheng
    Xu, Chao
    Rui, Yong
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 32 (CYCLE 2), 2014, 32 : 865 - 873
  • [32] Large-margin classification with multiple decision rules
    Kimes, Patrick K.
    Hayes, David Neil
    Marron, J. S.
    Liu, Yufeng
    STATISTICAL ANALYSIS AND DATA MINING, 2016, 9 (02) : 89 - 105
  • [33] STABILITY ENHANCED LARGE-MARGIN CLASSIFIER SELECTION
    Sun, Will Wei
    Cheng, Guang
    Liu, Yufeng
    STATISTICA SINICA, 2018, 28 (01) : 1 - 25
  • [34] Scalable Large-Margin Structured Learning: Theory and Algorithms
    Huang, Liang
    Zhao, Kai
    53RD ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 7TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2015), 2015, : 19 - 20
  • [35] Large-Margin Metric Learning for Constrained Partitioning Problems
    Lajugie, Remi
    Arlot, Sylvain
    Bach, Francis
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 32 (CYCLE 1), 2014, 32
  • [36] Efficient Private Algorithms for Learning Large-Margin Halfspaces
    Huy Le Nguyen
    Ullman, Jonathan
    Zakynthinou, Lydia
    ALGORITHMIC LEARNING THEORY, VOL 117, 2020, 117 : 704 - 724
  • [37] Double Additive Margin Softmax Loss for Face Recognition
    Zhou, Shengwei
    Chen, Caikou
    Han, Guojiang
    Hou, Xielian
    APPLIED SCIENCES-BASEL, 2020, 10 (01):
  • [38] Additive Margin Softmax with Center Loss for Face Recognition
    Jiang, Mingchao
    Yang, Zhenguo
    Liu, Wenyin
    Liu, Xiaochun
    PROCEEDINGS OF 2018 THE 2ND INTERNATIONAL CONFERENCE ON VIDEO AND IMAGE PROCESSING (ICVIP 2018), 2018, : 1 - 6
  • [39] A Geometric Perspective of Large-Margin Training of Gaussian Models
    Xiao, Lin
    Deng, Li
    IEEE SIGNAL PROCESSING MAGAZINE, 2010, 27 (06) : 118 - 123
  • [40] A flexible probabilistic framework for large-margin mixture of experts
    Archit Sharma
    Siddhartha Saxena
    Piyush Rai
    Machine Learning, 2019, 108 : 1369 - 1393