EFFECTS OF GENDER INFORMATION IN TEXT-INDEPENDENT AND TEXT-DEPENDENT SPEAKER VERIFICATION

被引:0
|
作者
Kanervisto, Anssi [1 ]
Vestman, Ville [1 ]
Sahidullah, Md [1 ]
Hautamaki, Ville [1 ]
Kinnunen, Tomi [1 ]
机构
[1] Univ Eastern Finland, Sch Comp, Kuopio, Finland
关键词
Speaker verification; gender dependent system; gender classification;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
It is well-known that for speaker recognition task, gender-dependent acoustic modeling performs better than gender-independent modeling. The practice is to use the gender ground-truth and to train gender-dependent models. However, such information is not necessarily available, especially if speakers are remotely enrolled. A way to overcome this is to use a gender classification system, which introduces an additional layer of uncertainty. To date, such uncertainty has not been studied. We implement two gender classifier systems and test them with two different corpora and speaker verification systems. We find that estimated gender information can improve speaker verification accuracy over gender-independent methods. Our detailed analysis suggests that gender estimation should have a sufficiently high accuracy to yield improvements in speaker verification performance.
引用
收藏
页码:5360 / 5364
页数:5
相关论文
共 50 条
  • [1] A Survey on Text-Dependent and Text-Independent Speaker Verification
    Tu, Youzhi
    Lin, Weiwei
    Mak, Man-Wai
    IEEE ACCESS, 2022, 10 : 99038 - 99049
  • [2] Improving Robustness of Speaker Verification by Fusion of Prompted Text-Dependent and Text-Independent Operation Modalities
    Mporas, Iosif
    Safavi, Saeid
    Sotudeh, Reza
    SPEECH AND COMPUTER, 2016, 9811 : 378 - 385
  • [3] VQ score normalisation for text-dependent and text-independent speaker recognition
    Finan, RA
    Sapeluk, AT
    Damper, RI
    AUDIO- AND VIDEO-BASED BIOMETRIC PERSON AUTHENTICATION, 1997, 1206 : 211 - 218
  • [4] SpeakerStew: Scaling to Many Languages with a Triaged Multilingual Text-Dependent and Text-Independent Speaker Verification System
    Chojnacka, Roza
    Pelecanos, Jason
    Wang, Quan
    Moreno, Ignacio Lopez
    INTERSPEECH 2021, 2021, : 1064 - 1068
  • [5] Text-dependent and text-independent speaker recognition of reverberant speech based on CNN
    El-Moneim, Samia Abd
    Sedik, Ahmed
    Nassar, M. A.
    El-Fishawy, Adel S.
    Sharshar, A. M.
    Hassan, Shaimaa E. A.
    Mahmoud, Adel Zaghloul
    Dessouky, Moawd I.
    El-Banby, Ghada M.
    El-Samie, Fathi E. Abd
    El-Rabaie, El-Sayed M.
    Neyazi, Badawi
    Seddeq, H. S.
    Ismail, Nabil A.
    Khalaf, Ashraf A. M.
    Elabyad, G. S. M.
    INTERNATIONAL JOURNAL OF SPEECH TECHNOLOGY, 2021, 24 (04) : 993 - 1006
  • [6] Text-dependent and text-independent speaker recognition of reverberant speech based on CNN
    Samia Abd El-Moneim
    Ahmed Sedik
    M. A. Nassar
    Adel S. El-Fishawy
    A. M. Sharshar
    Shaimaa E. A. Hassan
    Adel Zaghloul Mahmoud
    Moawd I. Dessouky
    Ghada M. El-Banby
    Fathi E. Abd El-Samie
    El-Sayed M. El-Rabaie
    Badawi Neyazi
    H. S. Seddeq
    Nabil A. Ismail
    Ashraf A. M. Khalaf
    G. S. M. Elabyad
    International Journal of Speech Technology, 2021, 24 : 993 - 1006
  • [7] EXPLOITING SEQUENCE INFORMATION FOR TEXT-DEPENDENT SPEAKER VERIFICATION
    Dey, Subhadeep
    Motlicek, Petr
    Madikeri, Srikanth
    Ferras, Marc
    2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2017, : 5370 - 5374
  • [8] A tutorial on text-independent speaker verification
    Bimbot, F. (bimbot@irisa.fr), 1600, Hindawi Publishing Corporation (2004):
  • [9] A tutorial on text-independent speaker verification
    Bimbot, F
    Bonastre, JF
    Fredouille, C
    Gravier, G
    Magrin-Chagnolleau, I
    Meignier, S
    Merlin, T
    Ortega-García, J
    Petrovska-Delacrétaz, D
    Reynolds, DA
    EURASIP JOURNAL ON APPLIED SIGNAL PROCESSING, 2004, 2004 (04) : 430 - 451
  • [10] Segment-level Effects of Gender, Nationality and Emotion Information on Text-independent Speaker Verification
    Li, Kai
    Akagi, Masato
    Wu, Yibo
    Dang, Jianwu
    INTERSPEECH 2020, 2020, : 2987 - 2991