Prediction of Various Backchannel Utterances Based on Multimodal Information

被引:0
|
作者
Onishi, Toshiki [1 ]
Azuma, Naoki [1 ]
Kinoshita, Shunichi [1 ]
Ishii, Ryo [2 ]
Fukayama, Atsushi [2 ]
Nakamura, Takao [2 ]
Miyata, Akihiro [1 ]
机构
[1] Nihon Univ, Tokyo, Japan
[2] NTT Corp, Yokohama, Kanagawa, Japan
来源
PROCEEDINGS OF THE 23RD ACM INTERNATIONAL CONFERENCE ON INTELLIGENT VIRTUAL AGENTS, IVA 2023 | 2023年
关键词
multimodal interaction; communication; backchannel; TURN-TAKING; JAPANESE; FEATURES; ENGLISH;
D O I
10.1145/3570945.3607298
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The listener's backchannels are an important part of dialogues. With appropriate backchannels, people are able to smoothly promote dialogues. Thus, backchannels are considered to be important in dialogues between not only humans but also humans and agents. Progress has been made in studying dialogue agents that perform natural affable dialogue. However, we have not clarified whether the listener's various backchannel types are predictable using the speaker's multimodal information. In this paper, we attempt to predict a listener's various backchannel types on the basis of the speaker's multimodal information in dialogues. First, we construct a dialogue corpus that consists of multimodal information of a speaker's utterances and a listener's backchannels. Second, we construct machine learning models to predict a listener's various backchannel types on the basis of a speaker's multimodal information. Our results suggest that our model was able to predict a listener's various backchannel types on the basis of a speaker's multimodal information.
引用
收藏
页数:4
相关论文
共 50 条
  • [31] Jointly Learning Agent and Lane Information for Multimodal Trajectory Prediction
    Wang, Jie
    Guo, Caili
    Guo, Minan
    Chen, Jiujiu
    Proceedings - International Conference on Pattern Recognition, 2022, 2022-August : 2921 - 2927
  • [32] Explaining (Sarcastic) Utterances to Enhance Affect Understanding in Multimodal Dialogues
    Kumar, Shivani
    Mondal, Ishani
    Akhtar, Md Shad
    Chakraborty, Tanmoy
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 11, 2023, : 12986 - 12994
  • [33] A multimodal cognitive approach to aid the conceptualization of Spanish utterances with 'se'
    Lopez-Ozieblo, Renia
    COGNITIVE LINGUISTICS, 2020, 31 (04) : 677 - 710
  • [34] Weather Visibility Prediction Based on Multimodal Fusion
    Zhang, Chuang
    Wu, Ming
    Chen, Jinyu
    Chen, Kaiyan
    Zhang, Chi
    Xie, Chao
    Huang, Bin
    He, Zichen
    IEEE ACCESS, 2019, 7 : 74776 - 74786
  • [35] Path-based Multimodal Trajectories Prediction
    Zhao, Ziqi
    Duan, Yiping
    Tao, Xiaoming
    2022 IEEE 96TH VEHICULAR TECHNOLOGY CONFERENCE (VTC2022-FALL), 2022,
  • [36] Daily Precipitation Prediction Based on Multimodal Representation
    Wu, Guohan
    Chen, Chi-Hua
    2023 IEEE INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS, ICCE, 2023,
  • [37] Corpus-Based Study on Backchannel "Yes"
    Xia Pengzheng
    Zhang Xiaoyu
    PROCEEDINGS OF THE FOURTH NORTHEAST ASIA INTERNATIONAL SYMPOSIUM ON LANGUAGE, LITERATURE AND TRANSLATION, 2015, 2015, : 145 - 151
  • [38] Extracting movie scenes based on multimodal information
    Li, Y
    Kuo, CCJ
    STORAGE AND RETRIEVAL FOR MEDIA DATABASES 2002, 2002, 4676 : 383 - 394
  • [39] Multimodal Sentiment Analysis Based on Information Bottleneck
    Cheng, Zichen
    Li, Yan
    Ge, Jiangwei
    Jiu, Mengfei
    Zhang, Jingwei
    Computer Engineering and Applications, 2024, 60 (02) : 137 - 146
  • [40] Semantic based information fusion in a multimodal interface
    Russ, G
    Sallans, B
    Hareter, H
    HCI '05: PROCEEDINGS OF THE 2005 INTERNATIONAL CONFERENCE ON HUMAN-COMPUTER INTERACTION, 2005, : 94 - 100