MSP-GEO Corpus: A Multimodal Database for Understanding Video-Learning Experience

被引:0
|
作者
Salman, Ali N. [1 ]
Wang, Ning [2 ]
Martinez-Lucas, Luz [1 ]
Vidal, Andrea [1 ]
Busso, Carlos [1 ]
机构
[1] Univ Texas Dallas, Dept Elect & Comp Engn, Richardson, TX 75083 USA
[2] Univ Texas Dallas, Dept Sustainable Earth Syst Sci, Richardson, TX USA
关键词
engagement; boredom; confusion; learning experiences; facial analysis; gaze; affective computing; e-learning; online learning; HIGHER-EDUCATION; ENGAGEMENT;
D O I
10.1145/3678957.3685737
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Video-based learning has become a popular, scalable, and effective approach for students to learn new skills. Many of the challenges for video-based learning can be addressed with machine learning models. However, the available datasets often lack the rich source of data that is needed to accurately predict students learning experiences and outcomes. To address this limitation, we introduce the MSP-GEO corpus, a new multimodal database that contains detailed demographic and educational data, recordings of the students and their screens, and meta-data about the lecture during the learning experience. The MSP-GEO corpus was collected using a quasi-experimental pre-test/post-test design. It consists of more than 39,600 seconds (11 hours) of continuous facial footage from 76 participants watching one of three experimental videos on the topic of fossil formation, resulting in over one million facial images. The data collected includes 21 gaze synchronization points, webcam and monitor recordings, and metadata for pauses, plays, and timeline navigation. Additionally, we annotated the recordings for engagement, boredom, and confusion using human evaluators. The MSP-GEO corpus has the potential to improve the accuracy of video-based learning outcomes and experience predictions, facilitate research on the psychological processes of video-based learning, inform the design of instructional videos, and advance the development of learning analytics methods.
引用
收藏
页码:488 / 497
页数:10
相关论文
共 5 条
  • [1] Understanding Flow Experience in Video Learning by Multimodal Data
    Wang, Yankai
    Chen, Bing
    Liu, Hongyan
    Hu, Zhiguo
    INTERNATIONAL JOURNAL OF HUMAN-COMPUTER INTERACTION, 2024, 40 (12) : 3144 - 3158
  • [2] Multimodal Learning toward Micro-Video Understanding
    Nie L.
    Liu M.
    Song X.
    Synthesis Lectures on Image, Video, and Multimedia Processing, 2019, 9 (04): : 1 - 186
  • [3] Neural Multimodal Cooperative Learning Toward Micro-Video Understanding
    Wei, Yinwei
    Wang, Xiang
    Guan, Weili
    Nie, Liqiang
    Lin, Zhouchen
    Chen, Baoquan
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 (29) : 1 - 14
  • [4] DeepQoE: A Multimodal Learning Framework for Video Quality of Experience (QoE) Prediction
    Zhang, Huaizheng
    Dong, Linsen
    Gao, Guanyu
    Hu, Han
    Wen, Yonggang
    Guan, Kyle
    IEEE TRANSACTIONS ON MULTIMEDIA, 2020, 22 (12) : 3210 - 3223
  • [5] Understanding public opinions on Chinese short video platform by multimodal sentiment analysis using deep learning-based techniques
    Shi, Wei
    Zhang, Jing
    He, Shaoyi
    KYBERNETES, 2023,