Multimodal Deep Reinforcement Learning for Visual Security of Virtual Reality Applications

被引:0
|
作者
Andam, Amine [1 ]
Bentahar, Jamal [2 ,3 ]
Hedabou, Mustapha [1 ]
机构
[1] Mohammed VI Polytech Univ, Sch Comp Sci, Ben Guerir 43150, Morocco
[2] Khalifa Univ, Res Ctr 6G, Abu Dhabi, U Arab Emirates
[3] Concordia Univ, Concordia Inst Informat Syst Engn, Montreal, PQ H3G 1M8, Canada
来源
IEEE INTERNET OF THINGS JOURNAL | 2024年 / 11卷 / 24期
基金
加拿大自然科学与工程研究理事会;
关键词
Security; Visualization; Avatars; Internet of Things; Web conferencing; Three-dimensional displays; Deep reinforcement learning; Deep reinforcement learning (DRL); multimodal neural network; output security; virtual reality (VR); SELECTIVE ATTENTION; DOMINANCE; ONSETS;
D O I
10.1109/JIOT.2024.3450686
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The rapid development of virtual reality (VR) technologies is bringing unprecedented immersive experiences and unusual digital content. Nevertheless, these advancements introduce new security challenges, especially in safeguarding the visual content displayed by VR devices like VR glasses and head-mounted displays. Most existing approaches for visual output security rely exclusively on numerical data, such as object attributes and overlook the need of visual information necessary for thorough VR protection. Moreover, these approaches typically assume a fixed size input, failing to address the dynamic nature of VR where the number of virtual items is constantly changing. This article presents a multimodal deep reinforcement learning (MMDRL) approach to secure the visual outputs in VR applications. We formalize a Markov decision process (MDP) framework for the MMDRL agent that integrates both numerical and image data into the state space to effectively mitigate visual threats. Furthermore, our MMDRL agent is engineered to handle data of varying sizes, which makes it more suitable for VR environments. Results from our experiments demonstrate the agent's ability to successfully counteract visual attacks, significantly outperforming previous approaches. The ablation study confirms the important role of image data in improving the agent's performance, highlighting the efficacy of our multimodal approach. In addition, we provide a video demonstration to showcase these results. Finally, we open-source our VR testbed and source code for further testing and benchmarking.
引用
收藏
页码:39890 / 39900
页数:11
相关论文
共 50 条
  • [31] Online Multimodal Transportation Planning using Deep Reinforcement Learning
    Farahani, Amirreza
    Genga, Laura
    Dijkman, Remco
    2021 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2021, : 1691 - 1698
  • [32] Multimodal information bottleneck for deep reinforcement learning with multiple sensors
    You, Bang
    Liu, Huaping
    NEURAL NETWORKS, 2024, 176
  • [33] Multimodal Biometrics Fusion Algorithm Using Deep Reinforcement Learning
    Huang, Quan
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2022, 2022
  • [34] The virtual security reality
    Franko, Matthew
    COMMUNICATIONS NEWS, 2008, 45 (04): : 39 - 39
  • [35] Deep Reinforcement Learning for Virtual Bidding in Electricity Markets
    Han D.
    Huang W.
    Yan Z.
    Zhongguo Dianji Gongcheng Xuebao/Proceedings of the Chinese Society of Electrical Engineering, 2022, 42 (04): : 1443 - 1454
  • [36] DeepViNE: Virtual Network Embedding with Deep Reinforcement Learning
    Dolati, Mahdi
    Hassanpour, Seyedeh Bahereh
    Ghaderi, Majid
    Khonsari, Ahmad
    IEEE CONFERENCE ON COMPUTER COMMUNICATIONS WORKSHOPS (IEEE INFOCOM 2019 WKSHPS), 2019, : 879 - 885
  • [37] Deep Reinforcement Learning for Task Planning of Virtual Characters
    Souza, Caio
    Velhor, Luiz
    INTELLIGENT COMPUTING, VOL 2, 2021, 284 : 694 - 711
  • [38] A Framework for the Development of Applications Allowing Multimodal Interaction with Virtual Reality Worlds
    Olmedo-Rodriguez, Hector
    Cardenoso-Payo, Valentin
    Escudero-Mancebo, David
    WSCG 2008, COMMUNICATION PAPERS, 2008, : 79 - 86
  • [39] A formal description of multimodal interaction techniques for immersive virtual reality applications
    Navarre, D
    Palanque, P
    Bastide, R
    Schyn, A
    Winckler, M
    Nedel, LP
    Freitas, CMDS
    HUMAN-COMPUTER INTERACTION - INTERACT 2005, PROCEEDINGS, 2005, 3585 : 170 - 183
  • [40] VIVID: Virtual Environment for Visual Deep Learning
    Lai, Kuan-Ting
    Lin, Chia-Chih
    Kang, Chun-Yao
    Liao, Mei-Enn
    Chen, Ming-Syan
    PROCEEDINGS OF THE 2018 ACM MULTIMEDIA CONFERENCE (MM'18), 2018, : 1356 - 1359