ZSTAD: Zero-Shot Temporal Activity Detection

被引:19
|
作者
Zhang, Lingling [1 ,2 ]
Chang, Xiaojun [3 ]
Liu, Jun [1 ,4 ]
Luo, Minnan [1 ,4 ]
Wang, Sen [5 ]
Ge, Zongyuan [3 ]
Hauptmann, Alexander [6 ]
机构
[1] Xi An Jiao Tong Univ, Sch Comp Sci & Technol, Xian, Peoples R China
[2] Minist Educ, Key Lab Intelligent Networks & Network Secur, Xian, Peoples R China
[3] Monash Univ, Fac Informat Technol, Melbourne, Vic, Australia
[4] Xi An Jiao Tong Univ, Natl Engn Lab Big Data Analyt, Xian, Peoples R China
[5] Univ Queensland, Sch Informat Technol & Elect Engn, Brisbane, Qld, Australia
[6] Carnegie Mellon Univ, Sch Comp Sci, Pittsburgh, PA 15213 USA
基金
澳大利亚研究理事会; 中国国家自然科学基金;
关键词
D O I
10.1109/CVPR42600.2020.00096
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
An integral part of video analysis and surveillance is temporal activity detection, which means to simultaneously recognize and localize activities in long untrimmed videos. Currently, the most effective methods of temporal activity detection are based on deep learning, and they typically perform very well with large scale annotated videos for training. However, these methods are limited in real applications due to the unavailable videos about certain activity classes and the time-consuming data annotation. To solve this challenging problem, we propose a novel task setting called zero-shot temporal activity detection (ZSTAD), where activities that have never been seen in training can still be detected. We design an end-to-end deep network based on R-C3D as the architecture for this solution. The proposed network is optimized with an innovative loss function that considers the embeddings of activity labels and their superclasses while learning the common semantics of seen and unseen activities. Experiments on both the THUMOS'14 and the Charades datasets show promising performance in terms of detecting unseen activities.
引用
收藏
页码:876 / 885
页数:10
相关论文
共 50 条
  • [1] TN-ZSTAD: Transferable Network for Zero-Shot Temporal Activity Detection
    Zhang, Lingling
    Chang, Xiaojun
    Liu, Jun
    Luo, Minnan
    Li, Zhihui
    Yao, Lina
    Hauptmann, Alex
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (03) : 3848 - 3861
  • [2] Semantics Guided Contrastive Learning of Transformers for Zero-shot Temporal Activity Detection
    Nag, Sayak
    Goldstein, Orpaz
    Roy-Chowdhury, Amit K.
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 6232 - 6242
  • [3] Zero-Shot Object Detection
    Bansal, Ankan
    Sikka, Karan
    Sharma, Gaurav
    Chellappa, Rama
    Divakaran, Ajay
    COMPUTER VISION - ECCV 2018, PT I, 2018, 11205 : 397 - 414
  • [4] ZERO-SHOT OBJECT DETECTION WITH TRANSFORMERS
    Zheng, Ye
    Cui, Li
    2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 444 - 448
  • [5] A Survey of Zero-Shot Object Detection
    Cao, Weipeng
    Yao, Xuyang
    Xu, Zhiwu
    Liu, Ye
    Pan, Yinghui
    Ming, Zhong
    BIG DATA MINING AND ANALYTICS, 2025, 8 (03): : 726 - 750
  • [6] A Survey of Zero-Shot Stance Detection
    Liu, Guangzhen
    Zhao, Kai
    Zhang, Linlin
    Bi, Xuehua
    Lv, Xiaoyi
    Chen, Cheng
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT V, NLPCC 2024, 2025, 15363 : 107 - 120
  • [7] Zero-Shot Camouflaged Object Detection
    Li, Haoran
    Feng, Chun-Mei
    Xu, Yong
    Zhou, Tao
    Yao, Lina
    Chang, Xiaojun
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 : 5126 - 5137
  • [8] Zero-Shot Defect Feature Optimizer: an efficient zero-shot optimization method for defect detection
    Yan, Zhibo
    Wu, Hanyang
    Aasim, Tehreem
    Yao, Haitao
    Zhang, Teng
    Wang, Dongyun
    JOURNAL OF ELECTRONIC IMAGING, 2025, 34 (01)
  • [9] Zero-Shot Temporal Action Detection via Vision-Language Prompting
    Nag, Sauradip
    Zhu, Xiatian
    Song, Yi-Zhe
    Xiang, Tao
    COMPUTER VISION - ECCV 2022, PT III, 2022, 13663 : 681 - 697
  • [10] Zero-Shot Object Detection for Indoor Robots
    Abdalwhab, Abdalwhab
    Liu, Huaping
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,