Data-Efficient Training Strategies for Neural TTS Systems

被引:2
|
作者
Prajwal, K. R. [1 ]
Jawahar, C., V [1 ]
机构
[1] IIIT Hyderabad, Hyderabad, India
关键词
transfer learning; multilingual training; text-to-speech;
D O I
10.1145/3430984.3431034
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
India is a country with thousands of languages and dialects spoken across a billion-strong population. For multi-lingual content creation and accessibility, text-to-speech systems will play a crucial role. However, the current neural TTS systems are data-hungry and need about 20 hours of clean single-speaker speech data for each language and speaker. This is not scalable for the large number of Indian languages and dialects. In this work, we demonstrate three simple, yet effective pre-training strategies that allow us to train neural TTS systems with just about one-tenth of the data needs while also achieving better accuracy and naturalness. We show that such pre-trained neural TTS systems can be quickly adapted to different speakers across languages and genders with less than 2 hours of data, thus significantly reducing the effort for future expansions to the thousands of rare Indian languages. We specifically highlight the benefits of multi-lingual pre-training and its consistent impact across our neural TTS systems for 8 Indian languages.
引用
收藏
页码:223 / 227
页数:5
相关论文
共 50 条
  • [1] Data-Efficient Augmentation for Training Neural Networks
    Liu, Tian Yu
    Mirzasoleiman, Baharan
    [J]. Advances in Neural Information Processing Systems, 2022, 35
  • [2] Data-Efficient Augmentation for Training Neural Networks
    Liu, Tian Yu
    Mirzasoleiman, Baharan
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [3] Sobolev Training for Data-efficient Approximate Nonlinear MPC
    Lueken, Lukas
    Brandner, Dean
    Lucia, Sergio
    [J]. IFAC PAPERSONLINE, 2023, 56 (02): : 5765 - 5772
  • [4] Data-efficient performance learning for configurable systems
    Jianmei Guo
    Dingyu Yang
    Norbert Siegmund
    Sven Apel
    Atrisha Sarkar
    Pavel Valov
    Krzysztof Czarnecki
    Andrzej Wasowski
    Huiqun Yu
    [J]. Empirical Software Engineering, 2018, 23 : 1826 - 1867
  • [5] A Data-Efficient Training Method for Deep Reinforcement Learning
    Feng, Wenhui
    Han, Chongzhao
    Lian, Feng
    Liu, Xia
    [J]. ELECTRONICS, 2022, 11 (24)
  • [6] Data-efficient performance learning for configurable systems
    Guo, Jianmei
    Yang, Dingyu
    Siegmund, Norbert
    Apel, Sven
    Sarkar, Atrisha
    Valov, Pavel
    Czarnecki, Krzysztof
    Wasowski, Andrzej
    Yu, Huiqun
    [J]. EMPIRICAL SOFTWARE ENGINEERING, 2018, 23 (03) : 1826 - 1867
  • [7] Intelligent Ultrasonic Systems for Material Texture Recognition using Data-Efficient Neural Networks
    Zhang, Xin
    Yu, Xinrui
    Saniie, Jafar
    [J]. INTERNATIONAL ULTRASONICS SYMPOSIUM (IEEE IUS 2021), 2021,
  • [8] Data-efficient Neural Text Compression with Interactive Learning
    Avinesh, P. V. S.
    Meyer, Christian M.
    [J]. 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 2543 - 2554
  • [9] ResMLP: Feedforward Networks for Image Classification With Data-Efficient Training
    Touvron, Hugo
    Bojanowski, Piotr
    Caron, Mathilde
    Cord, Matthieu
    El-Nouby, Alaaeldin
    Grave, Edouard
    Izacard, Gautier
    Joulin, Armand
    Synnaeve, Gabriel
    Verbeek, Jakob
    Jegou, Herve
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (04) : 5314 - 5321
  • [10] Training data-efficient image transformers & distillation through attention
    Touvron, Hugo
    Cord, Matthieu
    Douze, Matthijs
    Massa, Francisco
    Sablayrolles, Alexandre
    Jegou, Herve
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139 : 7358 - 7367