Multi-modal Human-Computer Virtual Fusion Interaction In Mixed Reality

被引:0
|
作者
Jia, Shengying [1 ]
机构
[1] Hulunbuir Univ, Hulunbuir 021000, Inner Mongolia, Peoples R China
来源
关键词
interactive visual perception; human-computer interaction; deep multi-modal fusion; mixed reality; INTERFACE;
D O I
10.6180/jase.202311_26(11).0010
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Since the receptive field of CNN usually reflects the size of its learning ability, it is limited by the size of the convolution kernel. At the same time, the use of pooling to increase the receptive field will cause the lack of spatial information of the feature map. While large receptive fields do not cause information loss, a deep multi -modal fusion line-of-sight tracking model based on dilated convolution is proposed. Using dilated convolution to further improve Res Net-50, and through experiments, it is proved that the use of dilated convolution can further improve the performance of the model. Comparing the designed gaze tracking model with the CNN-based gaze tracking model shows the results of the superiority of the gaze tracking model. In order to minimize the number of manual interventions, this paper adopts an adaptive target tracking method to achieve automatic collection of training samples. Based on the idea of active learning, the learning algorithm selects the sample containing the most information from the input stream of the training sample (the matching confidence given by the nearest neighbor classifier is lower than the set threshold) for constructing the perceptual model. A feature that is invariant to changes in rotation, brightness, and contrast is selected as the target descriptor to enhance the discriminative ability of the perceptual model. The experimental results verify the effectiveness of the multi-modal interactive visual perception method.
引用
收藏
页码:1609 / 1618
页数:10
相关论文
共 50 条
  • [1] An Immersive System with Multi-modal Human-computer Interaction
    Zhao, Rui
    Wang, Kang
    Divekar, Rahul
    Rouhani, Robert
    Su, Hui
    Ji, Qiang
    [J]. PROCEEDINGS 2018 13TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE & GESTURE RECOGNITION (FG 2018), 2018, : 517 - 524
  • [2] Experimental Study on Appropriate Reality of Agents as a Multi-modal Interface for Human-Computer Interaction
    Tanaka, Kaori
    Matsui, Tatsunori
    Kojima, Kazuaki
    [J]. HUMAN-COMPUTER INTERACTION: INTERACTION TECHNIQUES AND ENVIRONMENTS, PT II, 2011, 6762 : 613 - 622
  • [3] On Annotation and Evaluation of Multi-modal Corpora in Affective Human-Computer Interaction
    Kaechele, Markus
    Schels, Martin
    Meudt, Sascha
    Kessler, Viktor
    Glodek, Michael
    Thiam, Patrick
    Tschechne, Stephan
    Palm, Guenther
    Schwenker, Friedhelm
    [J]. MULTIMODAL ANALYSES ENABLING ARTIFICIAL AGENTS IN HUMAN-MACHINE INTERACTION, 2015, 8757 : 35 - 44
  • [4] An application of multi-modal human-computer interaction -: The chess player Turk 2
    Sajo, Levente
    Kovacs, Gyoergy
    Fazekas, Attila
    [J]. 2008 IEEE INTERNATIONAL CONFERENCE ON AUTOMATION, QUALITY AND TESTING, ROBOTICS (AQTR 2008), THETA 16TH EDITION, VOL II, PROCEEDINGS, 2008, : 316 - 319
  • [5] Human-computer interaction for alert warning and attention allocation systems of the Multi-Modal Watchstation
    Obermayer, RW
    Nugent, WA
    [J]. INTEGRATED COMMAND ENVIRONMENTS, 2000, 4126 : 14 - 22
  • [6] Realization of Human-computer Interaction Functions in Virtual Reality
    Zhong, Kunxia
    Zong, Meiling
    Guo, Zhengqin
    Wang, Zuolong
    [J]. 2009 INTERNATIONAL CONFERENCE ON RESEARCH CHALLENGES IN COMPUTER SCIENCE, ICRCCS 2009, 2009, : 229 - 231
  • [7] Virtual reality and multimedia human-computer interaction in medicine
    Englmeier, KH
    Krapichler, C
    Haubner, M
    Seemann, M
    Reiser, M
    [J]. 1998 IEEE SECOND WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING, 1998, : 193 - 202
  • [8] VIRTUAL REALITY BASED MULTI-MODAL TELEOPERATION USING MIXED AUTONOMY
    Kadavasal, Muthukkumar S.
    Seth, Abhishek
    Oliver, James H.
    [J]. DETC 2008: PROCEEDINGS OF THE ASME INTERNATIONAL DESIGN ENGINEERING TECHNICAL CONFERENCES AND COMPUTERS AND INFORMATIONAL IN ENGINEERING CONFERENCE, VOL 3, PTS A AND B: 28TH COMPUTERS AND INFORMATION IN ENGINEERING CONFERENCE, 2009, : 1451 - 1460
  • [9] Brain Signal Authentication for Human-computer Interaction in Virtual Reality
    Li, Sukun
    Savaliya, Sonal
    Marino, Leonard
    Leider, Avery M.
    Tappert, Charles C.
    [J]. 2019 22ND IEEE INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE AND ENGINEERING (IEEE CSE 2019) AND 17TH IEEE INTERNATIONAL CONFERENCE ON EMBEDDED AND UBIQUITOUS COMPUTING (IEEE EUC 2019), 2019, : 115 - 120
  • [10] Multi-modal hand gesture designing in multi-screen touchable teaching system for human-computer interaction
    Liu, Tingting
    Chen, Zengzhao
    Liu, Hai
    Zhang, Zhaoli
    Chen, Yingying
    [J]. ICAIP 2018: 2018 THE 2ND INTERNATIONAL CONFERENCE ON ADVANCES IN IMAGE PROCESSING, 2018, : 198 - 202