Social interaction of humanoid robot based on audio-visual tracking

被引:0
|
作者
Okuno, HG [1 ]
Nakadai, K
Kitano, H
机构
[1] Kyoto Univ, Grad Sch Informat, Kyoto 6068501, Japan
[2] ERATO, Japan Sci & Technol Corp, Kitano Symbiot Syst Project, Tokyo 1500001, Japan
[3] Sony Comp Sci Labs Inc, Shinagawa, Tokyo 1410022, Japan
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Social interaction is essential in improving robot human interface. Such behaviors for social interaction may include paying attention to a new sound source, moving toward it, or keeping face to face with a moving speaker. Some sound-centered behaviors may be difficult to attain, because the mixture of sounds is not well treated or auditory processing is too slow for real-time applications. Recently, Nakadai et al have developed real-time auditory and visual multiple-talker tracking technology by associating auditory and visual streams. The system is implemented on an upper-torso humanoid and the real-time talker tracking is attained with 200 msec of delay by distributed processing on four PCs connected by Gigabit Ethernet. Focus-of-attention is programmable and allows a variety of behaviors. The system demonstrates non-verbal social interaction by realizing a receptionist robot by focusing on an associated stream, while a companion robot on an auditory stream.
引用
收藏
页码:725 / 735
页数:11
相关论文
共 50 条
  • [1] AUDIO-VISUAL CONVERSATION ANALYSIS BY SMART POSTERBOARD AND HUMANOID ROBOT
    Kawahara, Tatsuya
    Inoue, Koji
    Lala, Divesh
    Takanashi, Katsuya
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 6573 - 6577
  • [2] Teaching a Humanoid Robot: Headset-Free Speech Interaction for Audio-Visual Association Learning
    Heckmann, Martin
    Brandl, Holger
    Schmuedderich, Jens
    Domont, Xavier
    Bolder, Bram
    Mikhailova, Inna
    Janssen, Herbert
    Gienger, Michael
    Bendig, Achim
    Rodemann, Tobias
    Dunn, Mark
    Joublin, Frank
    Goerick, Christian
    [J]. RO-MAN 2009: THE 18TH IEEE INTERNATIONAL SYMPOSIUM ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION, VOLS 1 AND 2, 2009, : 102 - 107
  • [3] Audio-visual Human Tracking for Active Robot Perception
    Bayram, Baris
    Ince, Gokhan
    [J]. 2015 23RD SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2015, : 1264 - 1267
  • [4] Audio-visual feedback improves the BCI performance in the navigational control of a humanoid robot
    Tidoni, Emmanuele
    Gergondet, Pierre
    Kheddar, Abderrahmane
    Aglioti, Salvatore M.
    [J]. FRONTIERS IN NEUROROBOTICS, 2014, 8
  • [5] Audio-Visual SLAM towards Human Tracking and Human-Robot Interaction in Indoor Environments
    Chau, Aaron
    Sekiguchi, Kouhei
    Nugraha, Aditya Arie
    Yoshii, Kazuyoshi
    Funakoshi, Kotaro
    [J]. 2019 28TH IEEE INTERNATIONAL CONFERENCE ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION (RO-MAN), 2019,
  • [6] Audio-Visual Bimodal Combination-Based Speaker Tracking Method for Mobile Robot
    Zhang, Hao-Yan
    Zhang, Long-Bo
    Shi, Qi-Feng
    Liu, Zhen-Tao
    [J]. JOURNAL OF ADVANCED COMPUTATIONAL INTELLIGENCE AND INTELLIGENT INFORMATICS, 2024, 28 (01) : 196 - 205
  • [7] Sporadic Audio-Visual Embodied Assistive Robot Navigation For Human Tracking
    Singh, Gaurav
    Ghanem, Paul
    Padir, Taskin
    [J]. PROCEEDINGS OF THE 16TH ACM INTERNATIONAL CONFERENCE ON PERVASIVE TECHNOLOGIES RELATED TO ASSISTIVE ENVIRONMENTS, PETRA 2023, 2023, : 99 - 105
  • [8] Audio-visual Multi-person Tracking for Active Robot Perception
    Bayram, Baris
    Ince, Gokhan
    [J]. 2015 IEEE/SICE INTERNATIONAL SYMPOSIUM ON SYSTEM INTEGRATION (SII), 2015, : 575 - 580
  • [9] Fast Audio-Visual Multi-Person Tracking for a Humanoid Stereo Camera Head
    Nickel, Kai
    Stiefethagen, Rainer
    [J]. HUMANOIDS: 2007 7TH IEEE-RAS INTERNATIONAL CONFERENCE ON HUMANOID ROBOTS, 2007, : 434 - 441
  • [10] Audio-Visual Sound Source Localization and Tracking Based on Mobile Robot for The Cocktail Party Problem
    Shi, Zhanbo
    Zhang, Lin
    Wang, Dongqing
    [J]. APPLIED SCIENCES-BASEL, 2023, 13 (10):