Enhancing learning outcomes through multisensory integration: A fMRI study of audio-visual training in virtual reality

被引:1
|
作者
Alwashmi, Kholoud [1 ,4 ,6 ]
Meyer, Georg [2 ]
Rowe, Fiona [3 ]
Ward, Ryan [2 ,5 ]
机构
[1] Univ Liverpool, Fac Hlth & Life Sci, Liverpool, England
[2] Univ Liverpool, Digital Innovat Facil, Liverpool, England
[3] Univ Liverpool, Inst Populat Hlth, Liverpool, England
[4] Princess Nourah bint Abdulrahman Univ, Dept Radiol, Riyadh, Saudi Arabia
[5] Liverpool John Moores Univ, Sch Comp Sci & Math, Liverpool, England
[6] Univ Liverpool, Eleanor Rathbone Bldg,Bedford St South, Liverpool L69 7ZA, England
关键词
fMRI; Multisensory; Audio-visual; Learning; Virtual; -reality; Eye; -movement; INFERIOR PARIETAL CORTEX; VISUAL-MOTION SIGNALS; SPATIAL ATTENTION; NEURAL RESPONSES; SPEECH SOUNDS; TIME-COURSE; PERFORMANCE; ACTIVATION; PLASTICITY; STIMULI;
D O I
10.1016/j.neuroimage.2023.120483
中图分类号
Q189 [神经科学];
学科分类号
071006 ;
摘要
The integration of information from different sensory modalities is a fundamental process that enhances perception and performance in real and virtual environments (VR). Understanding these mechanisms, especially during learning tasks that exploit novel multisensory cue combinations provides opportunities for the development of new rehabilitative interventions.This study aimed to investigate how functional brain changes support behavioural performance improvements during an audio-visual (AV) learning task. Twenty healthy participants underwent a 30 min daily VR training for four weeks. The task was an AV adaptation of a 'scanning training' paradigm that is commonly used in hemianopia rehabilitation. Functional magnetic resonance imaging (fMRI) and performance data were collected at baseline, after two and four weeks of training, and four weeks post-training. We show that behavioural performance, operationalised as mean reaction time reduction in VR, significantly improves. In separate tests in a controlled laboratory environment, we showed that the behavioural performance gains in the VR training environment transferred to a significant mean RT reduction for the trained AV voluntary task on a computer screen. Enhancements were observed in both the visual-only and AV conditions, with the latter demonstrating a faster response time supported by the presence of audio cues. The behavioural learning effect also transfers to two additional tasks that were tested: a visual search task and an involuntary visual task. Our fMRI results reveal an increase in functional activation (BOLD signal) in multisensory brain regions involved in early-stage AV processing: the thalamus, the caudal inferior parietal lobe and cerebellum. These functional changes were only observed for the trained, multisensory, task and not for unimodal visual stimulation. Functional activation changes in the thalamus were significantly correlated to behavioural performance improvements.This study demonstrates that incorporating spatial auditory cues to voluntary visual training in VR leads to augmented brain activation changes in multisensory integration, resulting in measurable performance gains across tasks. The findings highlight the potential of VR-based multisensory training as an effective method for enhancing cognitive function and as a potentially valuable tool in rehabilitative programmes.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] Neural correlates of audio-visual integration: an fMRI study
    Handley, R.
    Reinders, S.
    Marques, T.
    Pariante, C.
    McGuire, P.
    Dazzan, P.
    [J]. EARLY INTERVENTION IN PSYCHIATRY, 2008, 2 : A97 - A97
  • [2] Enhancing L2 sound learning through the integration of audio-visual information: Phonetic training in the classroom
    Li, Ying
    [J]. LANGUAGE TEACHING RESEARCH, 2024,
  • [3] Audio-visual saliency prediction with multisensory perception and integration
    Xie, Jiawei
    Liu, Zhi
    Li, Gongyang
    Song, Yingjie
    [J]. IMAGE AND VISION COMPUTING, 2024, 143
  • [4] Sensorimotor synchronization with audio-visual stimuli: limited multisensory integration
    Armstrong, Alan
    Issartel, Johann
    [J]. EXPERIMENTAL BRAIN RESEARCH, 2014, 232 (11) : 3453 - 3463
  • [5] Sensorimotor synchronization with audio-visual stimuli: limited multisensory integration
    Alan Armstrong
    Johann Issartel
    [J]. Experimental Brain Research, 2014, 232 : 3453 - 3463
  • [6] Detection of inconsistent audio-visual events in virtual reality
    Sorkin, A.
    Peled, A. ]
    Weinshall, D.
    [J]. PERCEPTION, 2006, 35 : 203 - 204
  • [7] Audio-visual integration through the parallel visual pathways
    Kaposvari, Peter
    Csete, Gergo
    Bognar, Anna
    Csibri, Peter
    Toth, Eszter
    Szabo, Nikoletta
    Vecsei, Laszlo
    Sary, Gyula
    Kincses, Zsigmond Tamas
    [J]. BRAIN RESEARCH, 2015, 1624 : 71 - 77
  • [8] An audio-visual multisensory training reduces interhemispheric imbalance in hemianopic patients
    Bertini, C.
    Pietrelli, M.
    Grasso, P. A.
    Zanon, M.
    Ladavas, E.
    [J]. JOURNAL OF THE NEUROLOGICAL SCIENCES, 2017, 381 : 602 - 602
  • [9] Audio-visual multisensory training enhances visual processing of motion stimuli in healthy participants: an electrophysiological study
    Grasso, Paolo A.
    Benassi, Mariagrazia
    Ladavas, Elisabetta
    Bertini, Caterina
    [J]. EUROPEAN JOURNAL OF NEUROSCIENCE, 2016, 44 (10) : 2748 - 2758
  • [10] Impact of Audio-Visual Complexity on Symptomatology of Laryngeal Dystonia: A Virtual Reality Study
    Petit, Jimmy
    Ehrlich, Stefan K.
    Tougas, Garrett
    Bernstein, Jacob M.
    Buie, Nicole E.
    Simonyan, Kristina
    [J]. LARYNGOSCOPE, 2024,