Annotating Objects and Relations in User-Generated Videos

被引:88
|
作者
Shang, Xindi [1 ]
Di, Donglin [1 ]
Xiao, Junbin [1 ]
Cao, Yu [1 ]
Yang, Xun [1 ]
Chua, Tat-Seng [1 ]
机构
[1] Natl Univ Singapore, Singapore, Singapore
基金
新加坡国家研究基金会;
关键词
dataset; video annotation; video content analysis; object recognition; visual relation recognition;
D O I
10.1145/3323873.3325056
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Understanding the objects and relations between them is indispensable to fine-grained video content analysis, which is widely studied in recent research works in multimedia and computer vision. However, existing works are limited to evaluating with either small datasets or indirect metrics, such as the performance over images. The underlying reason is that the construction of a large-scale video dataset with dense annotation is tricky and costly. In this paper, we address several main issues in annotating objects and relations in user-generated videos, and propose an annotation pipeline that can be executed at a modest cost. As a result, we present a new dataset, named VidOR, consisting of 10k videos (84 hours) together with dense annotations that localize 80 categories of objects and 50 categories of predicates in each video. We have made the training and validation set public and extendable for more tasks to facilitate future research on video object and relation recognition.
引用
收藏
页码:279 / 287
页数:9
相关论文
共 50 条
  • [1] ViComp: composition of user-generated videos
    Bano, Sophia
    Cavallaro, Andrea
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2016, 75 (12) : 7187 - 7210
  • [2] Multimodal Summarization of User-Generated Videos
    Psallidas, Theodoros
    Koromilas, Panagiotis
    Giannakopoulos, Theodoros
    Spyrou, Evaggelos
    [J]. APPLIED SCIENCES-BASEL, 2021, 11 (11):
  • [3] Predicting Emotions in User-Generated Videos
    Jiang, Yu-Gang
    Xu, Baohan
    Xue, Xiangyang
    [J]. PROCEEDINGS OF THE TWENTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2014, : 73 - 79
  • [4] ViComp: composition of user-generated videos
    Sophia Bano
    Andrea Cavallaro
    [J]. Multimedia Tools and Applications, 2016, 75 : 7187 - 7210
  • [5] A corpus of debunked and verified user-generated videos
    Papadopoulou, Olga
    Zampoglou, Markos
    Papadopoulos, Symeon
    Kompatsiaris, Ioannis
    [J]. ONLINE INFORMATION REVIEW, 2019, 43 (01) : 72 - 88
  • [6] Characterizing and modelling popularity of user-generated videos
    Borghol, Youmna
    Mitra, Siddharth
    Ardon, Sebastien
    Carlsson, Niklas
    Eager, Derek
    Mahanti, Anirban
    [J]. PERFORMANCE EVALUATION, 2011, 68 (11) : 1037 - 1055
  • [7] User-generated videos and tourists' intention to visit
    Adeloye, David
    Makurumidze, Kudzai
    Sarfo, Christian
    [J]. ANATOLIA-INTERNATIONAL JOURNAL OF TOURISM AND HOSPITALITY RESEARCH, 2022, 33 (04): : 658 - 671
  • [8] Revisiting Popularity Characterization and Modeling of User-generated Videos
    Islam, M. Aminul
    Eager, Derek
    Carlsson, Niklas
    Mahanti, Anirban
    [J]. 2013 IEEE 21ST INTERNATIONAL SYMPOSIUM ON MODELING, ANALYSIS & SIMULATION OF COMPUTER AND TELECOMMUNICATION SYSTEMS (MASCOTS 2013), 2013, : 350 - +
  • [9] On Evaluating Perceptual Quality of Online User-Generated Videos
    Jang, Soobeom
    Lee, Jong-Seok
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2016, 18 (09) : 1808 - 1818
  • [10] Recognition of Emotions in User-Generated Videos With Kernelized Features
    Zhang, Haimin
    Xu, Min
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2018, 20 (10) : 2824 - 2835