Memeplate: A Chinese Multimodal Dataset for Humor Understanding in Meme Templates

被引:0
|
作者
Li, Zefeng [1 ]
Lin, Hongfei [1 ]
Yang, Liang [1 ]
Xu, Bo [1 ]
Zhang, Shaowu [1 ]
机构
[1] Dalian Univ Technol, Dalian 116024, Peoples R China
基金
中国国家自然科学基金;
关键词
Multimodality; Sentiment analysis; Humor recognition;
D O I
10.1007/978-3-031-17120-8_41
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Humor plays an important role in human communication. Besides language, multimodal information is also of great significance in humor expression and understanding, which promotes the development of multimodal humor research. However, in existing datasets, images and text often have a one-to-one relationship, making it difficult to control image modality variables. It causes the low correlation and low enhancement between the two modalities in humor recognition tasks. Moreover, with the development of Vision Transformers (ViTs), the generalization ability of visual models has been greatly enhanced. Using ViTs alone can achieve impressive performance, but is difficult to explain. In this paper, we introduce Memeplate (Our dataset is available at https:// github.com/chineselzf/memeplate.), a novel multimodal humor dataset containing 203 templates, 5,184 memes and manually annotated humor levels. The template transfers images and text into a one-to-many relationship, which can make it easier for researchers to cut through the linguistic lens to multimodal humor. And it provides examples closer to human behavior for generation research. In addition, we provide multiple baseline results on the humor recognition task, which demonstrate the effectiveness of our control over image modality and the importance of introducing multimodal cues.
引用
收藏
页码:527 / 538
页数:12
相关论文
共 31 条
  • [31] Understanding public opinions on Chinese short video platform by multimodal sentiment analysis using deep learning-based techniques
    Shi, Wei
    Zhang, Jing
    He, Shaoyi
    [J]. KYBERNETES, 2023,