Protecting gender and identity with disentangled speech representations

被引:6
|
作者
Stoidis, Dimitrios [1 ]
Cavallaro, Andrea [1 ]
机构
[1] Queen Mary Univ London, Ctr Intelligent Sensing, London, England
来源
基金
英国工程与自然科学研究理事会;
关键词
privacy; soft biometrics; disentangled representation learning; variational autoencoder; IDENTIFICATION;
D O I
10.21437/Interspeech.2021-2163
中图分类号
R36 [病理学]; R76 [耳鼻咽喉科学];
学科分类号
100104 ; 100213 ;
摘要
Besides its linguistic content, our speech is rich in biometric information that can be inferred by classifiers. Learning privacy-preserving representations for speech signals enables downstream tasks without sharing unnecessary, private information about an individual. In this paper, we show that protecting gender information in speech is more effective than modelling speaker-identity information only when generating a nonsensitive representation of speech. Our method relies on reconstructing speech by decoding linguistic content along with gender information using a variational autoencoder. Specifically, we exploit disentangled representation learning to encode information about different attributes into separate subspaces that can be factorised independently. We present a novel way to encode gender information and disentangle two sensitive biometric identifiers, namely gender and identity, in a privacyprotecting setting. Experiments on the LibriSpeech dataset show that gender recognition and speaker verification can be reduced to a random guess, protecting against classification-based attacks.
引用
收藏
页码:1699 / 1703
页数:5
相关论文
共 50 条
  • [11] Signalling gender identity through speech
    Holmes, J
    [J]. MODERNA SPRAK, 1998, 92 (02): : 122 - 128
  • [12] Audio-guided self-supervised learning for disentangled visual speech representations
    Feng, Dalu
    Yang, Shuang
    Shan, Shiguang
    Chen, Xilin
    [J]. FRONTIERS OF COMPUTER SCIENCE, 2024, 18 (06)
  • [13] Adversarially Learning Disentangled Speech Representations for Robust Multi-factor Voice Conversion
    Wang, Jie
    Li, Jingbei
    Zhao, Xintao
    Wu, Zhiyong
    Kang, Shiyin
    Meng, Helen
    [J]. INTERSPEECH 2021, 2021, : 846 - 850
  • [14] Audio-guided self-supervised learning for disentangled visual speech representations
    FENG Dalu
    YANG Shuang
    SHAN Shiguang
    CHEN Xilin
    [J]. Frontiers of Computer Science., 2024, 18 (06)
  • [15] Learning Disentangled Representations for Recommendation
    Ma, Jianxin
    Zhou, Chang
    Cui, Peng
    Yang, Hongxia
    Zhu, Wenwu
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [16] Disentangled representations for causal cognition
    Torresan, Filippo
    Baltieri, Manuel
    [J]. Physics of Life Reviews, 2024, 51 : 343 - 381
  • [17] Social identity of gender and representations of human rights
    Jurberg, MB
    Nobrega, NP
    [J]. INTERNATIONAL JOURNAL OF PSYCHOLOGY, 1996, 31 (3-4) : 5197 - 5197
  • [18] Learning Disentangled Discrete Representations
    Friede, David
    Reimers, Christian
    Stuckenschmidt, Heiner
    Niepert, Mathias
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, ECML PKDD 2023, PT IV, 2023, 14172 : 593 - 609
  • [19] CONTRASTIVE PREDICTIVE CODING SUPPORTED FACTORIZED VARIATIONAL AUTOENCODER FOR UNSUPERVISED LEARNING OF DISENTANGLED SPEECH REPRESENTATIONS
    Ebbers, Janek
    Kuhlmann, Michael
    Cord-Landwehr, Tobias
    Haeb-Umbach, Reinhold
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 3860 - 3864
  • [20] Space and identity in the speech of Tunja: a gender study
    Bustamante Velez, Lucia
    [J]. CUADERNOS DE LINGUISTICA HISPANICA, 2015, 25 : 17 - 37