Weakly Supervised Video Salient Object Detection

被引:75
|
作者
Zhao, Wangbo [1 ]
Zhang, Jing [2 ,3 ]
Li, Long [1 ]
Barnes, Nick [2 ]
Liu, Nian [4 ]
Han, Junwei [1 ]
机构
[1] Northwestern Polytech Univ, Brain & Artificial Intelligence Lab, Xian, Peoples R China
[2] Australian Natl Univ, Canberra, ACT, Australia
[3] CSIRO, Canberra, ACT, Australia
[4] Incept Inst Artificial Intelligence, Abu Dhabi, U Arab Emirates
基金
美国国家科学基金会;
关键词
SEGMENTATION;
D O I
10.1109/CVPR46437.2021.01655
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Significant performance improvement has been achieved for fully-supervised video salient object detection with the pixel-wise labeled training datasets, which are time-consuming and expensive to obtain. To relieve the burden of data annotation, we present the first weakly supervised video salient object detection model based on relabeled "fixation guided scribble annotations". Specifically, an "Appearance-motion fusion module" and bidirectional ConvLSTM based framework are proposed to achieve effective multi-modal learning and long-term temporal context modeling based on our new weak annotations. Further, we design a novel foreground-background similarity loss to further explore the labeling similarity across frames. A weak annotation boosting strategy is also introduced to boost our model performance with a new pseudo-label generation technique. Extensive experimental results on six benchmark video saliency detection datasets illustrate the effectiveness of our solution(1).
引用
收藏
页码:16821 / 16830
页数:10
相关论文
共 50 条
  • [41] Weakly supervised salient object detection via bounding-box annotation and SAM model
    Liu, Xiangquan
    Huang, Xiaoming
    ELECTRONIC RESEARCH ARCHIVE, 2024, 32 (03): : 1624 - 1645
  • [42] WeakPCSOD: Overcoming the Bias of Box Annotations for Weakly Supervised Point Cloud Salient Object Detection
    Wei, Jun
    Zhou, S. Kevin
    Cui, Shuguang
    Li, Zhen
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 6, 2024, : 5894 - 5902
  • [43] Deep Hypersphere Feature Regularization for Weakly Supervised RGB-D Salient Object Detection
    Liu, Zhiyu
    Hayat, Munawar
    Yang, Hong
    Peng, Duo
    Lei, Yinjie
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 : 5423 - 5437
  • [44] Scribble-based complementary graph reasoning network for weakly supervised salient object detection
    Liang, Shuang
    Yan, Zhiqi
    Xie, Chi
    Zhu, Hongming
    Wang, Jiewen
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2024, 243
  • [45] Synthetic Data Supervised Salient Object Detection
    Wu, Zhenyu
    Wang, Lin
    Wang, Wei
    Shi, Tengfei
    Chen, Chenglizhao
    Hao, Aimin
    Li, Shuo
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 5557 - 5565
  • [46] Semi-Supervised Video Salient Object Detection Using Pseudo-Labels
    Yan, Pengxiang
    Li, Guanbin
    Xie, Yuan
    Li, Zhen
    Wang, Chuan
    Chen, Tianshui
    Lin, Liang
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 7283 - 7292
  • [47] Weakly Supervised Object Detection in Artworks
    Gonthier, Nicolas
    Gousseau, Yann
    Ladjal, Said
    Bonfait, Olivier
    COMPUTER VISION - ECCV 2018 WORKSHOPS, PT II, 2019, 11130 : 692 - 709
  • [48] Misclassification in Weakly Supervised Object Detection
    Wu, Zhihao
    Xu, Yong
    Yang, Jian
    Li, Xuelong
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 3413 - 3427
  • [49] Progressive Frame-Proposal Mining for Weakly Supervised Video Object Detection
    Han, Mingfei
    Wang, Yali
    Li, Mingjie
    Chang, Xiaojun
    Yang, Yi
    Qiao, Yu
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 1560 - 1573
  • [50] Object Instance Mining for Weakly Supervised Object Detection
    Lin, Chenhao
    Wang, Siwen
    Xu, Dongqi
    Lu, Yu
    Zhang, Wayne
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 11482 - 11489