The catchment feature model for multimodal language analysis

被引:0
|
作者
Quek, F [1 ]
机构
[1] Wright State Univ, Vis Interfaces & Sys Lab, Dayton, OH 45435 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Catchment Feature Model (CFM) addresses two questions in multimodal interaction: how do we bridge video and audio processing with the realities of human multimodal communication, and how information from the different modes may be fused. We discuss the need for our model, motivate the CFM from psycholinguistic research, and present the Model. In contrast to 'whole gesture' recognition, the CFM applies a feature decomposition approach that facilitates cross-modal fusion at the level of discourse planning and conceptualization. We present our experimental framework for CFM-based research, and cite three concrete examples of Catchment Features (CF), and propose new directions of multimodal research based on the model.
引用
收藏
页码:540 / 547
页数:8
相关论文
共 50 条
  • [1] The catchment feature model: A device for multimodal fusion and a bridge between signal and sense
    Quek, F
    [J]. EURASIP JOURNAL ON APPLIED SIGNAL PROCESSING, 2004, 2004 (11) : 1619 - 1636
  • [2] The Catchment Feature Model: A Device for Multimodal Fusion and a Bridge between Signal and Sense
    Francis Quek
    [J]. EURASIP Journal on Advances in Signal Processing, 2004
  • [3] A short video sentiment analysis model based on multimodal feature fusion
    Shi, Hongyu
    [J]. SYSTEMS AND SOFT COMPUTING, 2024, 6
  • [4] Feature Regression for Multimodal Image Analysis
    Yang, Michael Ying
    Yong, Xuanzi
    Rosenhahn, Bodo
    [J]. 2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2014, : 770 - 777
  • [5] Multimodal Discourse Analysis of Advertising Language
    Zhang, Chunli
    [J]. PROCEEDINGS OF 4TH INTERNATIONAL CONFERENCE ON EDUCATION, LANGUAGE, ART AND INTERCULTURAL COMMUNICATION (ICELAIC 2017), 2017, 142 : 169 - 172
  • [6] Comparative Analysis of Multimodal Large Language Model Performance on Clinical Vignette Questions
    Han, Tianyu
    Adams, Lisa C.
    Bressem, Keno K.
    Busch, Felix
    Nebelung, Sven
    Truhn, Daniel
    [J]. JAMA-JOURNAL OF THE AMERICAN MEDICAL ASSOCIATION, 2024, 331 (15): : 1320 - 1321
  • [7] Event detection using multimodal feature analysis
    Li, ZY
    Tan, YP
    [J]. 2005 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), VOLS 1-6, CONFERENCE PROCEEDINGS, 2005, : 3845 - 3848
  • [8] FMML: A Feature Model Markup Language
    Nabdel, Leili
    Karatas, Ahmet Serkan
    Oguztuzun, Halit
    Dogru, Ali
    [J]. NUMERICAL ANALYSIS AND APPLIED MATHEMATICS ICNAAM 2011: INTERNATIONAL CONFERENCE ON NUMERICAL ANALYSIS AND APPLIED MATHEMATICS, VOLS A-C, 2011, 1389
  • [9] Multimodal Feature Fusion Based Hypergraph Learning Model
    Yang, Zhe
    Xu, Liangkui
    Zhao, Lei
    [J]. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2022, 2022
  • [10] Multimodal Routing: Improving Local and Global Interpretability of Multimodal Language Analysis
    Tsai, Yao-Hung Hubert
    Ma, Martin Q.
    Yang, Muqiao
    Salakhutdinov, Ruslan
    Morency, Louis-Philippe
    [J]. PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 1823 - 1833