When Robots Get Chatty: Grounding Multimodal Human-Robot Conversation and Collaboration

被引:0
|
作者
Allgeuer, Philipp [1 ]
Ali, Hassan [1 ]
Wermter, Stefan [1 ]
机构
[1] Univ Hamburg, Dept Informat, Knowledge Technol, Hamburg, Germany
关键词
Natural Dialog for Robots; LLM Grounding; AI-Enabled Robotics; Multimodal Interaction;
D O I
10.1007/978-3-031-72341-4_21
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We investigate the use of Large Language Models (LLMs) to equip neural robotic agents with human-like social and cognitive competencies, for the purpose of open-ended human-robot conversation and collaboration. We introduce a modular and extensible methodology for grounding an LLM with the sensory perceptions and capabilities of a physical robot, and integrate multiple deep learning models throughout the architecture in a form of system integration. The integrated models encompass various functions such as speech recognition, speech generation, open-vocabulary object detection, human pose estimation, and gesture detection, with the LLM serving as the central text-based coordinating unit. The qualitative and quantitative results demonstrate the huge potential of LLMs in providing emergent cognition and interactive language-oriented control of robots in a natural and social manner. Video: https://youtu.be/A2WLEuiM3-s.
引用
收藏
页码:306 / 321
页数:16
相关论文
共 50 条
  • [31] A multimodal human-robot sign language interaction framework applied in social robots
    Li, Jie
    Zhong, Junpei
    Wang, Ning
    FRONTIERS IN NEUROSCIENCE, 2023, 17
  • [32] Research on Human-Robot Collaboration Method for Parallel Robots Oriented to Segment Docking
    Sun, Deyuan
    Wang, Junyi
    Xu, Zhigang
    Bao, Jianwen
    Lu, Han
    SENSORS, 2024, 24 (06)
  • [33] A Method Based on Wearable Devices for Controlling Teaching of Robots for Human-robot Collaboration
    Tao, Yong
    Fang, Zengliang
    Ren, Fan
    Wang, Tianmiao
    Deng, Xianling
    Sun, Baishu
    2018 CHINESE AUTOMATION CONGRESS (CAC), 2018, : 2270 - 2276
  • [34] Multimodal perception-fusion-control and human-robot collaboration in manufacturing: a review
    Duan, Jianguo
    Zhuang, Liwen
    Zhang, Qinglei
    Zhou, Ying
    Qin, Jiyun
    INTERNATIONAL JOURNAL OF ADVANCED MANUFACTURING TECHNOLOGY, 2024, 132 (3-4): : 1071 - 1093
  • [35] Nonverbal Signaling for Non-Humanoid Robots During Human-Robot Collaboration
    Cha, Elizabeth
    Mataric, Maja
    Fong, Terrence
    ELEVENTH ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN ROBOT INTERACTION (HRI'16), 2016, : 601 - 602
  • [36] Deep Learning-based Multimodal Control Interface for Human-Robot Collaboration
    Liu, Hongyi
    Fang, Tongtong
    Zhou, Tianyu
    Wang, Yuquan
    Wang, Lihui
    51ST CIRP CONFERENCE ON MANUFACTURING SYSTEMS, 2018, 72 : 3 - 8
  • [37] A Self-Modulated Impedance Multimodal Interaction Framework for Human-Robot Collaboration
    Muratore, Luca
    Laurenzi, Arturo
    Tsagarakis, Nikos G.
    2019 INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2019, : 4998 - 5004
  • [38] The RoSiD Tool: Empowering Users to Design Multimodal Signals for Human-Robot Collaboration
    Dennler, Nathaniel
    Delgado, David
    Zeng, Daniel
    Nikolaidis, Stefanos
    Mataric, Maja
    EXPERIMENTAL ROBOTICS, ISER 2023, 2024, 30 : 3 - 10
  • [39] Impact of Robot Initiative on Human-Robot Collaboration
    Munzer, Thibaut
    Mollard, Yoan
    Lopes, Manuel
    COMPANION OF THE 2017 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION (HRI'17), 2017, : 217 - 218
  • [40] Robust Robot Planning for Human-Robot Collaboration
    You, Yang
    Thomas, Vincent
    Colas, Francis
    Alami, Rachid
    Buffet, Olivier
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2023), 2023, : 9793 - 9799