A Speech-Centric Perspective for Human-Computer Interface: A Case Study

被引:0
|
作者
Li Deng
Dong Yu
机构
[1] Microsoft Research,
关键词
human-computer interaction; speech-centric multimodal interface; robust speech processing; MapPointS; speech-driven mobile navigation system;
D O I
暂无
中图分类号
学科分类号
摘要
Speech technology has been playing a central role in enhancing human-machine interactions, especially for small devices for which graphical user interface has obvious limitations. The speech-centric perspective for human-computer interface advanced in this paper derives from the view that speech is the only natural and expressive modality to enable people to access information from and to interact with any device. In this paper, we describe some recent work conducted at Microsoft Research, aimed at the development of enabling technologies for speech-centric multimodal human-computer interaction. In particular, we present a case study of a prototype system, called MapPointS, which is a speech-centric multimodal map-query application for North America. This prototype navigation system provides rich functionalities that allow users to obtain map-related information through speech, text, and pointing devices. Users can verbally query for state maps, city maps, directions, places, nearby businesses and other useful information within North America. They can also verbally control applications such as changing the map size and panning the map moving interactively through speech. In the current system, the results of the queries are presented back to users through graphical user interface. An overview and major components of the MapPointS system will be presented in detail first. This will be followed by software design engineering principles and considerations adopted in developing the MapPointS system, and by a description of some key robust speech processing technologies underlying general speech-centric human-computer interaction systems.
引用
收藏
页码:255 / 269
页数:14
相关论文
共 50 条
  • [1] A speech-centric perspective for human-computer interface: A case study
    Deng, L
    Yu, D
    [J]. JOURNAL OF VLSI SIGNAL PROCESSING SYSTEMS FOR SIGNAL IMAGE AND VIDEO TECHNOLOGY, 2005, 41 (03): : 255 - 269
  • [2] A speech-centric perspective for human-computer interface
    Deng, L
    Acero, A
    Wang, Y
    Wang, K
    Hon, H
    Droppo, J
    Mahajan, M
    Huang, XD
    [J]. PROCEEDINGS OF THE 2002 IEEE WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING, 2002, : 263 - 267
  • [3] Speech recognition in the human-computer interface
    Rebman, CM
    Aiken, MW
    Cegielski, CG
    [J]. INFORMATION & MANAGEMENT, 2003, 40 (06) : 509 - 519
  • [4] Design of an Embedded Speech-Centric Interface for Applications in Handheld Terminals
    Gallardo-Antolin, Ascension
    Garcia-Moral, Ana I.
    Pereiro-Estevan, Yago
    Diaz-de-Maria, Fernando
    [J]. IEEE AEROSPACE AND ELECTRONIC SYSTEMS MAGAZINE, 2013, 28 (02) : 24 - 33
  • [5] Effects of Stereoscopy on a Human-Computer Interface for Network Centric Operations
    Zocco, Alessandro
    Cannone, Davide
    De Paolis, Lucio Tommaso
    [J]. PROCEEDINGS OF THE 2014 9TH INTERNATIONAL CONFERENCE ON COMPUTER VISION, THEORY AND APPLICATIONS (VISAPP 2014), VOL 3, 2014, : 249 - 255
  • [6] Context-Centric Speech-Based Human-Computer Interaction
    Hung, Victor C.
    Gonzalez, Avelino J.
    [J]. INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2013, 28 (10) : 1010 - 1037
  • [7] Enhanced Human-Computer Speech Interface Using Wavelet Computing
    Ayat, Saeed
    [J]. 2008 IEEE INTERNATIONAL CONFERENCE ON VIRTUAL ENVIRONMENTS, HUMAN-COMPUTER INTERFACES AND MEASUREMENT SYSTEMS, 2008, : 37 - 40
  • [8] Human-computer interactions in speech therapy using a blowing interface
    Ruminski, Jacek
    Bujnowski, Adam
    Wtorek, Jerzy
    [J]. 2014 7TH INTERNATIONAL CONFERENCE ON HUMAN SYSTEM INTERACTIONS (HSI), 2014, : 178 - 181
  • [9] AN EXPERIMENTAL-STUDY OF THE HUMAN-COMPUTER INTERFACE
    BENBASAT, I
    DEXTER, AS
    MASULIS, PS
    [J]. COMMUNICATIONS OF THE ACM, 1981, 24 (11) : 752 - 762