Integrated multimodal human-computer interface and augmented reality for interactive display applications

被引:1
|
作者
Vassiliou, MS [1 ]
Sundareswaran, V [1 ]
Chen, S [1 ]
Behringer, R [1 ]
Tam, C [1 ]
Chan, M [1 ]
Bangayan, P [1 ]
McGee, J [1 ]
机构
[1] Rockwell Int Sci Ctr, Thousand Oaks, CA 91360 USA
关键词
human-computer interface; speech recognition; 3D Audio; eyetracking; multimodal integration; lip reading; multimedia; augmented reality; tactical operation center; wearable computing;
D O I
10.1117/12.397779
中图分类号
V [航空、航天];
学科分类号
08 ; 0825 ;
摘要
We describe new systems for improved integrated multimodal human-computer interaction and augmented reality for a diverse array of applications, including future advanced cockpits, tactical operations centers, and others. We have developed an integrated display system featuring: speech recognition of multiple concurrent users equipped with both standard air-coupled microphones and novel throat-coupled sensors (developed at Army Research Labs for increased noise immunity); lip reading for improving speech recognition accuracy in noisy environments, three-dimensional spatialized audio for improved display of warnings, alerts, and other information; wireless, coordinated handheld-PC control of a large display; real-time display of data and inferences from wireless integrated networked sensors with on-board signal processing and discrimination; gesture control with disambiguated point-and-speak capability; head- and eye-tracking coupled with speech recognition for "look-and-speak" interaction; and integrated tetherless augmented reality on a wearable computer. The various interaction modalities (speech recognition, 3D audio, eyetracking, etc.) are implemented as "modality servers" in an Internet-based client-server architecture. Each modality server encapsulates and exposes commercial and research software packages, presenting a socket network interface that is abstracted to a high-level interface, minimizing both vendor dependencies and required changes on the client side as the server's technology improves.
引用
收藏
页码:106 / 115
页数:10
相关论文
共 50 条
  • [1] Interactive Human-Computer Theoretical Model of Editorial Design with Augmented Reality
    Borja-Galeas, Carlos
    Guevara, Cesar
    [J]. INTELLIGENT HUMAN SYSTEMS INTEGRATION 2021, 2021, 1322 : 580 - 585
  • [2] Toward multimodal human-computer interface
    Sharma, R
    Pavlovic, VI
    Huang, TS
    [J]. PROCEEDINGS OF THE IEEE, 1998, 86 (05) : 853 - 869
  • [3] An augmented reality human-computer interface for object localization in a cognitive vision system
    Siegl, H.
    Hanheide, M.
    Wrede, S.
    Pinz, A.
    [J]. IMAGE AND VISION COMPUTING, 2007, 25 (12) : 1895 - 1903
  • [4] A novel interactive mouse system for holistic haptic display in a human-computer interface
    Kyung, Ki-Uk
    Kwon, Dong-Soo
    Yang, Gii-Hun
    [J]. INTERNATIONAL JOURNAL OF HUMAN-COMPUTER INTERACTION, 2006, 20 (03) : 247 - 270
  • [5] Study of Human-Computer Interaction in Augmented Reality
    Janani, S. K.
    Swarnalatha, P.
    [J]. SOFT COMPUTING FOR PROBLEM SOLVING, SOCPROS 2018, VOL 2, 2020, 1057 : 835 - 846
  • [6] An Human-Computer Interactive Augmented Reality System for Coronary Artery Diagnosis Planning and Training
    Qiming Li
    Chen Huang
    Shengqing Lv
    Zeyu Li
    Yimin Chen
    Lizhuang Ma
    [J]. Journal of Medical Systems, 2017, 41
  • [7] An Human-Computer Interactive Augmented Reality System for Coronary Artery Diagnosis Planning and Training
    Li, Qiming
    Huang, Chen
    Lv, Shengqing
    Li, Zeyu
    Chen, Yimin
    Ma, Lizhuang
    [J]. JOURNAL OF MEDICAL SYSTEMS, 2017, 41 (10)
  • [8] Augmented reality (AR): The blurring of reality in human-computer interaction
    Mahr, Dominik
    Heller, Jonas
    de Ruyter, Ko
    [J]. COMPUTERS IN HUMAN BEHAVIOR, 2023, 145
  • [9] Human-computer multimodal interface to internet navigation
    Mosquera-DeLaCruz, Jose H.
    Loaiza-Correa, Humberto
    Nope-Rodriguez, Sandra E.
    Restrepo-Giro, Andres D.
    [J]. DISABILITY AND REHABILITATION-ASSISTIVE TECHNOLOGY, 2021, 16 (08) : 807 - 820
  • [10] New Applications of Multimodal Human-Computer Interfaces
    Czyzewski, Andrzej
    [J]. 2012 JOINT CONFERENCE NEW TRENDS IN AUDIO & VIDEO AND SIGNAL PROCESSING: ALGORITHMS, ARCHITECTURES, ARRANGEMENTS, & APPLICATIONS (NTAV-SPA 2012), 2012, : 19 - 24