Web-based environment for user generation of spoken dialog for virtual assistants

被引:3
|
作者
Nishimura, Ryota [1 ]
Yamamoto, Daisuke [2 ]
Uchiya, Takahiro [2 ]
Takumi, Ichi [2 ]
机构
[1] Tokushima Univ, Dept Technol Ind & Social Sci, Tokushima, Japan
[2] Nagoya Inst Technol, Dept Comp Sci, Nagoya, Aichi, Japan
基金
日本科学技术振兴机构;
关键词
Spoken dialog system; MMDAgent; Digital signage; Real field; Web service; User-generated content; Consumer-generated media;
D O I
10.1186/s13636-018-0142-8
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
In this paper, a web-based spoken dialog generation environment which enables users to edit dialogs with a video virtual assistant is developed and to also select the 3D motions and tone of voice for the assistant. In our proposed system, anyone can easily post/edit contents of the dialog for the dialog system. The dialog type corresponding to the system is limited to the question-and-answer type dialog, in order to avoid editing conflicts caused by editing by multiple users. The spoken dialog sharing service and FST generator generates spoken dialog content for the MMDAgent spoken dialog system toolkit, which includes a speech recognizer, a dialog control unit, a speech synthesizer, and a virtual agent. For dialog content creation, question-and-answer dialogs posted by users and FST templates are used. The proposed system was operated for more than a year in a student lounge at the Nagoya Institute of Technology, where users added more than 500 dialogs during the experiment. Images were also registered to 65% of the postings. The most posted category is related to animation, video games, manga.The system was subjected to open examination by tourist information staff who had no prior experience with spoken dialog systems. Based on their impressions of tourist use of the dialog system, they shortened the length of some of the system's responses and added pauses to the longer responses to make them easier to understand.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] Web-based environment for user generation of spoken dialog for virtual assistants
    Ryota Nishimura
    Daisuke Yamamoto
    Takahiro Uchiya
    Ichi Takumi
    [J]. EURASIP Journal on Audio, Speech, and Music Processing, 2018
  • [2] The web-based domain independent DM for intelligent spoken dialog system
    Yuan, CX
    Wang, JH
    Liu, JY
    Cong, W
    [J]. 2005 IEEE NETWORKING, SENSING AND CONTROL PROCEEDINGS, 2005, : 1027 - 1030
  • [3] A Web-based virtual environment for operator training
    Tam, EK
    Badra, F
    Marceau, RJ
    Marin, MA
    Malowany, AS
    [J]. IEEE TRANSACTIONS ON POWER SYSTEMS, 1999, 14 (03) : 802 - 808
  • [4] Web-based configuration assistants
    Attardi, G
    Cisternino, A
    Simi, M
    [J]. AI EDAM-ARTIFICIAL INTELLIGENCE FOR ENGINEERING DESIGN ANALYSIS AND MANUFACTURING, 1998, 12 (04): : 321 - 331
  • [5] User behaviour in web-based interactive virtual tours
    Bastanlar, Yalin
    [J]. PROCEEDINGS OF THE ITI 2007 29TH INTERNATIONAL CONFERENCE ON INFORMATION TECHNOLOGY INTERFACES, 2007, : 221 - 226
  • [6] Creation of web-based user interface for supercomputing environment
    Wu, H
    Chi, XB
    Xu, F
    [J]. FIFTH INTERNATIONAL CONFERENCE ON ALGORITHMS AND ARCHITECTURES FOR PARALLEL PROCESSING, PROCEEDINGS, 2002, : 446 - 452
  • [7] A web-based virtual reality environment for medical visualization
    Kokelj, Ziga
    Bohak, Ciril
    Marolt, Matija
    [J]. 2018 41ST INTERNATIONAL CONVENTION ON INFORMATION AND COMMUNICATION TECHNOLOGY, ELECTRONICS AND MICROELECTRONICS (MIPRO), 2018, : 299 - 302
  • [8] CyberWalk: A web-based distributed virtual walkthrough environment
    Chim, J
    Lau, RWH
    Leong, HV
    Si, A
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2003, 5 (04) : 503 - 515
  • [9] Development of Web-based Virtual Training Environment for Machining
    Yang, Zhixin
    Wong, S. F.
    [J]. ISCM II AND EPMESC XII, PTS 1 AND 2, 2010, 1233 : 1285 - 1290
  • [10] Distributed virtual learning environment: a web-based approach
    Bouras, C
    Philopoulos, A
    [J]. PROCEEDINGS OF THE 26TH EUROMICRO CONFERENCE, VOLS I AND II, 2000, : A50 - A55