Triple Correlations-Guided Label Supplementation for Unbiased Video Scene Graph Generation

被引:1
|
作者
Wang, Wenqing [1 ]
Gao, Kaifeng [1 ]
Luo, Yawei [1 ]
Jiang, Tao [1 ]
Gao, Fei [2 ]
Shao, Jian [1 ]
Sun, Jianwen [3 ]
Xiao, Jun [1 ]
机构
[1] Zhejiang Univ, Hangzhou, Peoples R China
[2] Zhejiang Univ Technol, Hangzhou, Peoples R China
[3] Cent China Normal Univ, Wuhan, Peoples R China
基金
中国国家自然科学基金;
关键词
video scene graph generation; spatio-temporal correlations; long-tail problem; missing label supplementation;
D O I
10.1145/3581783.3612024
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Video-based scene graph generation (VidSGG) is an approach that aims to represent video content in a dynamic graph by identifying visual entities and their relationships. Due to the inherently biased distribution and missing annotations in the training data, current VidSGG methods have been found to perform poorly on less-represented predicates. In this paper, we propose an explicit solution to address this under-explored issue by supplementing missing predicates that should appear in the ground-truth annotations. Dubbed Trico, our method seeks to supplement the missing predicates by exploring three complementary spatio-temporal correlations. Guided by these correlations, the missing labels can be effectively supplemented thus achieving an unbiased predicate predictions. We validate the effectiveness of Trico on the most widely used VidSGG datasets, i.e., VidVRD and VidOR. Extensive experiments demonstrate the state-of-the-art performance achieved by Trico, particularly on those tail predicates. The code is available in https://github.com/Wq23333/Trico.git.
引用
收藏
页码:5153 / 5163
页数:11
相关论文
共 50 条
  • [1] Label Semantic Knowledge Distillation for Unbiased Scene Graph Generation
    Li, Lin
    Xiao, Jun
    Shi, Hanrong
    Wang, Wenxiao
    Shao, Jian
    Liu, An-An
    Yang, Yi
    Chen, Long
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (01) : 195 - 206
  • [2] Taking a Closer Look At Visual Relation: Unbiased Video Scene Graph Generation With Decoupled Label Learning
    Wang, Wenqing
    Luo, Yawei
    Chen, Zhiqing
    Jiang, Tao
    Yang, Yi
    Xiao, Jun
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 5718 - 5728
  • [3] TEMPLATE-GUIDED DATA AUGMENTATION FOR UNBIASED SCENE GRAPH GENERATION
    Zang, Yujie
    Li, Yaochen
    Cao, Luguang
    Lu, Ruitao
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024, 2024, : 3510 - 3514
  • [4] Unbiased Scene Graph Generation in Videos
    Nag, Sayak
    Min, Kyle
    Tripathi, Subama
    Roy-Chowdhury, Amit K.
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 22803 - 22813
  • [5] Adaptive Feature Learning for Unbiased Scene Graph Generation
    Yang, Jiarui
    Wang, Chuan
    Yang, Liang
    Jiang, Yuchen
    Cao, Angelina
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 2252 - 2265
  • [6] Unbiased Scene Graph Generation from Biased Training
    Tang, Kaihua
    Niu, Yulei
    Huang, Jianqiang
    Shi, Jiaxin
    Zhang, Hanwang
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 3713 - 3722
  • [7] Unbiased Scene Graph Generation Using Predicate Similarities
    Matsui, Yusuke
    Ohashi, Misaki
    IEEE ACCESS, 2024, 12 : 95507 - 95516
  • [8] Compositional Feature Augmentation for Unbiased Scene Graph Generation
    Li, Lin
    Chen, Guikun
    Xiao, Jun
    Yang, Yi
    Wang, Chunping
    Chen, Long
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 21628 - 21638
  • [9] Vision Relation Transformer for Unbiased Scene Graph Generation
    Sudhakaran, Gopika
    Dhami, Devendra Singh
    Kersting, Kristian
    Roth, Stefan
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 21825 - 21836
  • [10] Geometric and Semantic Improvement for Unbiased Scene Graph Generation
    Zhang, Ruhui
    Xu, Pengcheng
    Kang, Kang
    Yang, You
    KSII TRANSACTIONS ON INTERNET AND INFORMATION SYSTEMS, 2023, 17 (10): : 2643 - 2657