Speaker Verification Employing Combinations of Self-Attention Mechanisms

被引:5
|
作者
Bae, Ara [1 ]
Kim, Wooil [1 ]
机构
[1] Incheon Natl Univ, Dept Comp Sci & Engn, Incheon 22012, South Korea
关键词
speaker verification; self-attention; attention combinations;
D O I
10.3390/electronics9122201
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
One of the most recent speaker recognition methods that demonstrates outstanding performance in noisy environments involves extracting the speaker embedding using attention mechanism instead of average or statistics pooling. In the attention method, the speaker recognition performance is improved by employing multiple heads rather than a single head. In this paper, we propose advanced methods to extract a new embedding by compensating for the disadvantages of the single-head and multi-head attention methods. The combination method comprising single-head and split-based multi-head attentions shows a 5.39% Equal Error Rate (EER). When the single-head and projection-based multi-head attention methods are combined, the speaker recognition performance improves by 4.45%, which is the best performance in this work. Our experimental results demonstrate that the attention mechanism reflects the speaker's properties more effectively than average or statistics pooling, and the speaker verification system could be further improved by employing combinations of different attention techniques.
引用
收藏
页码:1 / 11
页数:11
相关论文
共 50 条
  • [21] SELF-SUPERVISED SPEAKER VERIFICATION EMPLOYING A NOVEL CLUSTERING ALGORITHM
    Fathan, Abderrahim
    Alam, Jahangir
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2024), 2024, : 12597 - 12601
  • [22] Self-attention based speaker recognition using Cluster-Range Loss
    Bian, Tengyue
    Chen, Fangzhou
    Xu, Li
    NEUROCOMPUTING, 2019, 368 : 59 - 68
  • [23] An innovative deep learning framework for skin cancer detection employing ConvNeXtV2 and focal self-attention mechanisms
    Ozdemir, Burhanettin
    Pacal, Ishak
    RESULTS IN ENGINEERING, 2025, 25
  • [24] SHYNESS AND SELF-ATTENTION
    CROZIER, WR
    BULLETIN OF THE BRITISH PSYCHOLOGICAL SOCIETY, 1983, 36 (FEB): : A5 - A5
  • [25] Aggregating Frame-Level Information in the Spectral Domain With Self-Attention for Speaker Embedding
    Tu, Youzhi
    Mak, Man-Wai
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2022, 30 : 944 - 957
  • [26] Attention and self-attention in random forests
    Utkin, Lev V.
    Konstantinov, Andrei V.
    Kirpichenko, Stanislav R.
    PROGRESS IN ARTIFICIAL INTELLIGENCE, 2023, 12 (03) : 257 - 273
  • [27] Attention and self-attention in random forests
    Lev V. Utkin
    Andrei V. Konstantinov
    Stanislav R. Kirpichenko
    Progress in Artificial Intelligence, 2023, 12 : 257 - 273
  • [28] GRAPH ATTENTION NETWORKS FOR SPEAKER VERIFICATION
    Jung, Jee-weon
    Heo, Hee-Soo
    Yu, Ha-Jin
    Chung, Joon Son
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 6149 - 6153
  • [29] Polarimetric Thermal to Visible Face Verification via Self-Attention Guided Synthesis
    Di, Xing
    Riggan, Benjamin S.
    Hu, Shuowen
    Short, Nathaniel J.
    Patel, Vishal M.
    2019 INTERNATIONAL CONFERENCE ON BIOMETRICS (ICB), 2019,
  • [30] The Contrastive Network With Convolution and Self-Attention Mechanisms for Unsupervised Cell Segmentation
    Zhao, Yuhang
    Shao, Xianhao
    Chen, Cai
    Song, Junlin
    Tian, Chongxuan
    Li, Wei
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2023, 27 (12) : 5837 - 5847