A Coarse-to-Fine Framework for Resource Efficient Video Recognition

被引:8
|
作者
Wu, Zuxuan [1 ]
Li, Hengduo [2 ]
Zheng, Yingbin [3 ]
Xiong, Caiming [4 ]
Jiang, Yu-Gang [1 ]
Davis, Larry S. [2 ]
机构
[1] Fudan Univ, Sch Comp Sci, Shanghai Key Lab Intelligent Informat Proc, Shanghai, Peoples R China
[2] Univ Maryland, College Pk, MD 20742 USA
[3] Videt Lab, Shanghai, Peoples R China
[4] Salesforce Res, Palo Alto, CA USA
基金
中国国家自然科学基金;
关键词
Conditional computation; Video classification; Efficient recognition;
D O I
10.1007/s11263-021-01508-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks have demonstrated remarkable recognition results on video classification, however great improvements in accuracies come at the expense of large amounts of computational resources. In this paper, we introduce LiteEval for resource efficient video recognition. LiteEval is a coarse-to-fine framework that dynamically allocates computation on a per-video basis, and can be deployed in both online and offline settings. Operating by default on low-cost features that are computed with images at a coarse scale, LiteEval adaptively determines on-the-fly when to read in more discriminative yet computationally expensive features. This is achieved by the interactions of a coarse RNN and a fine RNN, together with a conditional gating module that automatically learns when to use more computation conditioned on incoming frames. We conduct extensive experiments on three large-scale video benchmarks, FCVID, ActivityNet and Kinetics, and demonstrate, among other things, that LiteEval offers impressive recognition performance while using significantly less computation for both online and offline settings.
引用
收藏
页码:2965 / 2977
页数:13
相关论文
共 50 条
  • [1] A Coarse-to-Fine Framework for Resource Efficient Video Recognition
    Zuxuan Wu
    Hengduo Li
    Yingbin Zheng
    Caiming Xiong
    Yu-Gang Jiang
    Larry S Davis
    [J]. International Journal of Computer Vision, 2021, 129 : 2965 - 2977
  • [2] LiteEval: A Coarse-to-Fine Framework for Resource Efficient Video Recognition
    Wu, Zuxuan
    Xiong, Caiming
    Jiang, Yu-Gang
    Davis, Larry S.
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [3] A Coarse-to-Fine Framework for Automatic Video Unscreen
    Rao, Anyi
    Xu, Linning
    Li, Zhizhong
    Huang, Qingqiu
    Kuang, Zhanghui
    Zhang, Wayne
    Lin, Dahua
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 2723 - 2733
  • [4] CONE: An Efficient COarse-to-fiNE Alignment Framework for Long Video Temporal Grounding
    Hou, Zhijian
    Zhong, Wanjun
    Ji, Lei
    Gao, Difei
    Yan, Kun
    Chan, Wing-Kwong
    Ngo, Chong-Wah
    Shou, Mike Zheng
    Duan, Nan
    [J]. PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 8013 - 8028
  • [5] A COARSE-TO-FINE LOGO RECOGNITION METHOD IN VIDEO STREAMS
    Zhao, Chaoyang
    Wang, Jinqiao
    Xie, Chengli
    Lu, Hanqing
    [J]. 2014 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO WORKSHOPS (ICMEW), 2014,
  • [6] COARSE-TO-FINE VIDEO TEXT DETECTION
    Miao, Guangyi
    Huang, Qingming
    Jiang, Shuqiang
    Gao, Wen
    [J]. 2008 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, VOLS 1-4, 2008, : 569 - +
  • [7] A coarse-to-fine method for shape recognition
    Tang H.-X.
    Wei H.
    [J]. Journal of Computer Science and Technology, 2007, 22 (2) : 330 - 334
  • [8] A Coarse-to-Fine Deep Learning Based Framework for Traffic Light Recognition
    Yao, Zikai
    Liu, Qiang
    Fu, Jie
    Xie, Qian
    Li, Bo
    Ye, Qing
    Li, Qing
    [J]. IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024,
  • [9] A coarse-to-fine method for shape recognition
    Tang, Hui-Xuan
    Wei, Hui
    [J]. JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2007, 22 (02) : 329 - 333
  • [10] A Coarse-to-Fine Method for Data-Efficient Collaborative Place Recognition
    Furuno, Eike
    Hein, Andreas
    Stratmann, Tim C.
    Pfingsthorn, Max
    [J]. 2024 9TH INTERNATIONAL CONFERENCE ON CONTROL AND ROBOTICS ENGINEERING, ICCRE 2024, 2024, : 140 - 146