Machine learning methods for speech emotion recognition on telecommunication systemsMachine learning methods for speech emotion recognition on telecommunication systemsA. Osipov et al.

被引:0
|
作者
Alexey Osipov [1 ]
Ekaterina Pleshakova [1 ]
Yang Liu [2 ]
Sergey Gataullin [3 ]
机构
[1] MIREA - Russian Technological University,
[2] Xidian University,undefined
[3] Moscow Technical University of Communications and Informatics,undefined
关键词
Artificial intelligence; Neural networks; Engineering; CapsNet; Smart bracelet; Photoplethysmogram; Speech emotion recognition;
D O I
10.1007/s11416-023-00500-2
中图分类号
学科分类号
摘要
The manuscript is devoted to the study of human behavior in stressful situations using machine learning methods, which depends on the psychotype, socialization and a host of other factors. Global mobile subscribers lost approximately $53 billion in 2022 due to phone fraud and unwanted calls, with almost half (43%) of subscribers having spam blocking or caller ID apps installed. Phone scammers build their conversation focusing on the behavior of a certain category of people. Previously, a person is introduced into a state of acute stress, in which his further behavior to one degree or another can be manipulated. We were allowed to single out the target audience by research by Juniper Research. These are men under the age of 44 who have the highest risk of being deceived by scammers. This significantly narrows the scope of research and allows us to limit the behavioral features of this particular category of subscribers. In addition, this category of people uses modern gadgets, which allows researchers not to consider outdated models; has stable health indicators, which allows not to conduct additional studies of people with diseases of the heart system, because. Their percentage in this sample is minimal; and also most often undergoes a polygraph interview, for example, when applying for a job, and this allows us to get a sample sufficient for training the neural network. To teach the method, polygrams were used, marked by a polygraph examiner and a psychologist of healthy young people who underwent a scheduled polygraph test for company loyalty. For testing, the readings of the PPG sensor built into the smart bracelet were taken and analyzed within a month from young people who underwent a polygraph test. We have developed a modification of the wavelets capsular neural network—2D-CapsNet, allowing to identify the state of panic stupor by classification quality indicators: Accuracy—86.0%, Precision—84.0%, Recall = 87.5% and F-score—85.7%, according to the photoplethysmogram graph (PPG), which does not allow him to make logically sound decisions. When synchronizing a smart bracelet with a smartphone, the method allows real-time tracking of such states, which makes it possible to respond to a call from a telephone scammer during a conversation with a subscriber. The proposed method can be widely used in cyber-physical systems in order to detect illegal actions.
引用
收藏
页码:415 / 428
页数:13
相关论文
共 50 条
  • [31] Speech emotion recognition via learning analogies
    Ntalampiras, Stavros
    PATTERN RECOGNITION LETTERS, 2021, 144 : 21 - 26
  • [32] Discriminative Feature Learning for Speech Emotion Recognition
    Zhang, Yuying
    Zou, Yuexian
    Peng, Junyi
    Luo, Danqing
    Huang, Dongyan
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2019: TEXT AND TIME SERIES, PT IV, 2019, 11730 : 198 - 210
  • [33] Speech Emotion Recognition Using Deep Learning
    Alagusundari, N.
    Anuradha, R.
    ARTIFICIAL INTELLIGENCE: THEORY AND APPLICATIONS, VOL 1, AITA 2023, 2024, 843 : 313 - 325
  • [34] Speech Emotion Recognition Using Deep Learning
    Ahmed, Waqar
    Riaz, Sana
    Iftikhar, Khunsa
    Konur, Savas
    ARTIFICIAL INTELLIGENCE XL, AI 2023, 2023, 14381 : 191 - 197
  • [35] Active Learning for Dimensional Speech Emotion Recognition
    Han, Wenjing
    Li, Haifeng
    Ruan, Huabin
    Ma, Lin
    Sun, Jiayin
    Schuller, Bjoern
    14TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2013), VOLS 1-5, 2013, : 2840 - 2844
  • [36] Speech Emotion Recognition Using Transfer Learning
    Song, Peng
    Jin, Yun
    Zhao, Li
    Xin, Minghai
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2014, E97D (09): : 2530 - 2532
  • [37] Autonomous Emotion Learning in Speech: A View of Zero-Shot Speech Emotion Recognition
    Xu, Xinzhou
    Deng, Jun
    Cummins, Nicholas
    Zhang, Zixing
    Zhao, Li
    Schuller, Bjorn W.
    INTERSPEECH 2019, 2019, : 949 - 953
  • [38] Machine Learning Approaches for Speech Emotion Recognition: Classic and Novel Advances
    Heracleous, Panikos
    Ishikawa, Akio
    Yasuda, Keiji
    Kawashima, Hiroyuki
    Sugaya, Fumiaki
    Hashimoto, Masayuki
    COMPUTATIONAL LINGUISTICS AND INTELLIGENT TEXT PROCESSING, CICLING 2017, PT II, 2018, 10762 : 180 - 191
  • [39] SER: Speech Emotion Recognition Application Based on Extreme Learning Machine
    Ainurrochman
    Febriansyah, Irfanur Ilham
    Yuhana, Umi Laili
    PROCEEDINGS OF 2021 13TH INTERNATIONAL CONFERENCE ON INFORMATION & COMMUNICATION TECHNOLOGY AND SYSTEM (ICTS), 2021, : 179 - 183
  • [40] Speech emotion recognition method in educational scene based on machine learning
    Zhang, Yanning
    Srivastava, Gautam
    EAI ENDORSED TRANSACTIONS ON SCALABLE INFORMATION SYSTEMS, 2022, 9 (05)