Interactive Video Indexing With Statistical Active Learning

被引:101
|
作者
Zha, Zheng-Jun [1 ]
Wang, Meng [2 ]
Zheng, Yan-Tao [3 ]
Yang, Yi [4 ]
Hong, Richang [2 ]
Chua, Tat-Seng [1 ]
机构
[1] Natl Univ Singapore, Sch Comp, Singapore 117417, Singapore
[2] Hefei Univ Technol, Hefei, Anhui, Peoples R China
[3] Inst Infocomm Res, Singapore, Singapore
[4] Univ Queensland, Sch Informat Technol & Elect Engn, Brisbane, Qld, Australia
关键词
Active learning; optimum experimental design; video indexing; FRAMEWORK;
D O I
10.1109/TMM.2011.2174782
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Video indexing, also called video concept detection, has attracted increasing attentions from both academia and industry. To reduce human labeling cost, active learning has been introduced to video indexing recently. In this paper, we propose a novel active learning approach based on the optimum experimental design criteria in statistics. Different from existing optimum experimental design, our approach simultaneously exploits sample's local structure, and sample relevance, density, and diversity information, as well as makes use of labeled and unlabeled data. Specifically, we develop a local learning model to exploit the local structure of each sample. Our assumption is that for each sample, its label can be well estimated based on its neighbors. By globally aligning the local models from all the samples, we obtain a local learning regularizer, based on which a local learning regularized least square model is proposed. Finally, a unified sample selection approach is developed for interactive video indexing, which takes into account the sample relevance, density and diversity information, and sample efficacy in minimizing the parameter variance of the proposed local learning regularized least square model. We compare the performance between our approach and the state-of-the-art approaches on the TREC video retrieval evaluation (TRECVID) benchmark. We report superior performance from the proposed approach.
引用
下载
收藏
页码:17 / 27
页数:11
相关论文
共 50 条
  • [31] Deep Learning Based Semantic Video Indexing and Retrieval
    Podlesnaya, Anna
    Podlesnyy, Sergey
    PROCEEDINGS OF SAI INTELLIGENT SYSTEMS CONFERENCE (INTELLISYS) 2016, VOL 2, 2018, 16 : 359 - 372
  • [32] Self-supervised learning for robust video indexing
    Ewerth, Ralph
    Freisleben, Bernd
    2006 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO - ICME 2006, VOLS 1-5, PROCEEDINGS, 2006, : 1749 - +
  • [33] Active learning with multiple classifiers for multimedia indexing
    Bahjat Safadi
    Georges Quénot
    Multimedia Tools and Applications, 2012, 60 : 403 - 417
  • [34] Active learning with multiple classifiers for multimedia indexing
    Safadi, Bahjat
    Quenot, Georges
    MULTIMEDIA TOOLS AND APPLICATIONS, 2012, 60 (02) : 403 - 417
  • [35] Adding Interactive Mode to Active Video Watching
    Bojnordi, Ehsan
    Mitrovic, Antonija
    Galster, Matthias
    Malinen, Sanna
    Holland, Jay
    31ST INTERNATIONAL CONFERENCE ON COMPUTERS IN EDUCATION, ICCE 2023, VOL II, 2023, : 1042 - 1044
  • [36] Novel motion-based active video indexing method
    Wei, Jie
    Li, Ze-Nian
    Gertner, Izidor
    International Conference on Multimedia Computing and Systems -Proceedings, 1999, 2 : 460 - 465
  • [37] A novel motion-based active video indexing method
    Wei, J
    Li, ZN
    Gertner, I
    IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA COMPUTING AND SYSTEMS, PROCEEDINGS VOL 2, 1999, : 460 - 465
  • [38] Time-constrained dynamic semantic compression for video indexing and interactive searching
    Liu, TC
    Kender, JR
    2001 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOL 2, PROCEEDINGS, 2001, : 531 - 538
  • [39] Time-constrained dynamic semantic compression for video indexing and interactive searching
    Department of Computer Science, Columbia University, New York, NY 10027, United States
    Proc IEEE Comput Soc Conf Comput Vision Pattern Recognit, (II531-II538):
  • [40] Video for Active and Remote Learning
    Lipomi, Darren J.
    TRENDS IN CHEMISTRY, 2020, 2 (06): : 483 - 485