An interactive system to automatically generate video summaries and perform subtitles synchronization for persons with hearing loss

被引:0
|
作者
Cuzco-Calle, I. [1 ]
Ingavelez-Guerra, P. [1 ]
Robles-Bykbaev, V. [1 ]
Calle-Lopez, D. [1 ]
机构
[1] Catedra UNESCO Tecnol Apoyo Inclus Educat, GI IATa, Calle Vieja 12-30 & Elia Liut, Cuenca, Ecuador
关键词
Video summaries; Hearing loss; Disabilities; Latent Semantic Analysis;
D O I
暂无
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
According to World Health Organization (WHO), approximately 328 million of adults and 32 million of children present hearing loss in the world. Likewise, the number of people with such impairment increased from 42 million in 1985 to about 360 million in 2011. However, the most of multimedia and web contents, whether they are educational or leisure, are not accessible for persons with hearing loss. In this line, this paper presents an interactive system aimed at automatically generating video summaries and performing subtitles synchronization for persons with hearing loss. Our proposal relies on an educational platform (MOODLE) and Natural Language Processing (NLP) to provide an environment fully configurable for these persons. The module that generates the video summaries uses techniques such as Latent Semantic Analysis (LSA) and Latent Dirichlet Allocation (LDA), whereas the synchronization module is based on forced alignment between audio streams and text. With the aim of validating our environment, we have tested our approach on 15 videos, obtaining a score of 80% for three criteria related to the summary content: understandability, concordance, and context appropriateness.
引用
收藏
页数:4
相关论文
共 3 条
  • [1] An Intelligent System to Automatically Generate Video-Summaries for Accessible Learning Objects for People with Hearing Loss
    Ingavelez-Guerra, Paola
    Cuzco-Calle, Irma
    Calle-Lopez, Daniel
    Oyola-Flores, Christian
    Yambay-Aulla, Ines
    Robles-Bykbaev, Vladimir
    Ramon Hilera, Jose
    [J]. ADVANCES IN HUMAN FACTORS IN TRAINING, EDUCATION, AND LEARNING SCIENCES, AHFE 2017, 2018, 596 : 113 - 122
  • [2] A Method for Automatically Creating an Interactive Semantic Video based on AR System
    Lee, Kee-Sung
    Rosli, Ahmad Nurzid
    Jo, Geun-Sik
    [J]. PROCEEDINGS 2012 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2012, : 2626 - 2631
  • [3] Producing text and speech from video images of lips movement photographed in speaking Japanese by using mouth shape sequence code - An experimental system to communicate with hearing impaired persons -
    Kawahata, Shiori
    Koyama, Eiko
    Miyazaki, Tsuyoshi
    Yamamoto, Fujio
    [J]. PROCEEDINGS OF THE SEVENTEENTH INTERNATIONAL SYMPOSIUM ON ARTIFICIAL LIFE AND ROBOTICS (AROB 17TH '12), 2012, : 867 - 870