A gaze-based virtual keyboard using a mouth switch for command selection

被引:0
|
作者
Soundarajan, S. [1 ]
Cecotti, H. [1 ]
机构
[1] Fresno State Univ, Dept Comp Sci, Coll Sci & Math, Fresno, CA 93740 USA
关键词
D O I
暂无
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
Portable eye-trackers provide an efficient way to access the point of gaze from a user on a computer screen. Thanks to eyetracking, gaze-based virtual keyboard can be developed by taking into account constraints related to the gaze detection accuracy. In this paper, we propose a new gaze-based virtual keyboard where all the letters can be accessed directly through a single command. In addition, we propose a USB mouth switch that is directly connected through a computer mouse, with the mouse switch replacing the left click button. This approach is considered to tackle the Midas touch problem with eye-tracking for people who are severely disabled. The performance is evaluated on 10 participants by comparing the following three conditions: gaze detection with mouth switch, gaze detection with dwell time by considering the distance to the closest command, and the gaze detection within the surface of the command box. Finally, a workload using NASA-TLX test was conducted on the different conditions. The results revealed that the proposed approach with the mouth switch provides a better performance in terms of typing speed (36.6 +/- 8.4 letters/minute) compared to the other conditions, and a high acceptance as an input device.
引用
收藏
页码:3334 / 3337
页数:4
相关论文
共 50 条
  • [41] Leap to the Eye: Implicit Gaze-based Interaction to Reveal Invisible Objects for Virtual Environment Exploration
    Chen, Yang-Sheng
    Hsieh, Chiao-En
    Jie, Miguel Then Ying
    Han, Ping-Hsuan
    Hung, Yi-Ping
    2023 IEEE INTERNATIONAL SYMPOSIUM ON MIXED AND AUGMENTED REALITY, ISMAR, 2023, : 214 - 222
  • [42] Towards a Unified and Efficient Command Selection Mechanism for Touch-Based Devices Using Soft Keyboard Hotkeys
    Fennedy, Katherine
    Srivastava, Angad
    Malacria, Sylvain
    Perrault, Simon T.
    ACM TRANSACTIONS ON COMPUTER-HUMAN INTERACTION, 2022, 29 (01)
  • [43] Multi-stage gaze-controlled virtual keyboard using eye tracking
    Tatinyuy, Verdzekov Emile
    Woguia, Auguste Vigny Noumsi
    Ngono, Joseph mvogo
    Fono, Louis Aime
    PLOS ONE, 2024, 19 (10):
  • [44] EOG-based eye movement detection and gaze estimation for an asynchronous virtual keyboard
    Barbara, Nathaniel
    Camilleri, Tracey A.
    Camilleri, Kenneth P.
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2019, 47 : 159 - 167
  • [45] Predicting Gaze-based Target Selection in Augmented Reality Headsets based on Eye and Head Endpoint Distributions
    Wei, Yushi
    Shi, Rongkai
    Yu, Difeng
    Wang, Yihong
    Li, Yue
    Yu, Lingyun
    Liang, Hai-Ning
    PROCEEDINGS OF THE 2023 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS, CHI 2023, 2023,
  • [46] Optimal Gaze-Based Robot Selection in Multi-Human Multi-Robot Interaction
    Zhang, Lingkang
    Vaughan, Richard
    ELEVENTH ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN ROBOT INTERACTION (HRI'16), 2016, : 645 - 646
  • [47] Enhancing patient freedom in rehabilitation robotics using gaze-based intention detection
    Novak, Domen
    Riener, Robert
    2013 IEEE 13TH INTERNATIONAL CONFERENCE ON REHABILITATION ROBOTICS (ICORR), 2013,
  • [48] GazeCast: Using Mobile Devices to Allow Gaze-based Interaction on Public Displays
    Namnakani, Omar
    Sinrattanavong, Penpicha
    Abdrabou, Yasmeen
    Bulling, Andreas
    Alt, Florian
    Khamis, Mohamed
    ACM SYMPOSIUM ON EYE TRACKING RESEARCH & APPLICATIONS, ETRA 2023, 2023,
  • [49] Dwell time preferences for gaze-based object selection of different object types vary with age
    Remijn, Gerard
    Paulus, Yesaya Tommy
    PERCEPTION, 2022, 51 : 155 - 155
  • [50] Selection-based virtual keyboard prototypes and data collection application
    Barbara Millet
    Shihab Asfour
    James R. Lewis
    Behavior Research Methods, 2009, 41 : 951 - 956