Triple Correlations-Guided Label Supplementation for Unbiased Video Scene Graph Generation

被引:1
|
作者
Wang, Wenqing [1 ]
Gao, Kaifeng [1 ]
Luo, Yawei [1 ]
Jiang, Tao [1 ]
Gao, Fei [2 ]
Shao, Jian [1 ]
Sun, Jianwen [3 ]
Xiao, Jun [1 ]
机构
[1] Zhejiang Univ, Hangzhou, Peoples R China
[2] Zhejiang Univ Technol, Hangzhou, Peoples R China
[3] Cent China Normal Univ, Wuhan, Peoples R China
基金
中国国家自然科学基金;
关键词
video scene graph generation; spatio-temporal correlations; long-tail problem; missing label supplementation;
D O I
10.1145/3581783.3612024
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Video-based scene graph generation (VidSGG) is an approach that aims to represent video content in a dynamic graph by identifying visual entities and their relationships. Due to the inherently biased distribution and missing annotations in the training data, current VidSGG methods have been found to perform poorly on less-represented predicates. In this paper, we propose an explicit solution to address this under-explored issue by supplementing missing predicates that should appear in the ground-truth annotations. Dubbed Trico, our method seeks to supplement the missing predicates by exploring three complementary spatio-temporal correlations. Guided by these correlations, the missing labels can be effectively supplemented thus achieving an unbiased predicate predictions. We validate the effectiveness of Trico on the most widely used VidSGG datasets, i.e., VidVRD and VidOR. Extensive experiments demonstrate the state-of-the-art performance achieved by Trico, particularly on those tail predicates. The code is available in https://github.com/Wq23333/Trico.git.
引用
收藏
页码:5153 / 5163
页数:11
相关论文
共 50 条
  • [41] A New Training Data Organization Form and Training Mode for Unbiased Scene Graph Generation
    Xu, Hongbo
    Wang, Lichun
    Xu, Kai
    Fu, Fangyu
    Yin, Baocai
    Huang, Qingming
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (07) : 5295 - 5305
  • [42] HiLo: Exploiting High Low Frequency Relations for Unbiased Panoptic Scene Graph Generation
    Zhou, Zijian
    Shi, Miaojing
    Caesar, Holger
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 21580 - 21591
  • [43] Stacked Hybrid-Attention and Group Collaborative Learning for Unbiased Scene Graph Generation
    Dong, Xingning
    Gan, Tian
    Song, Xuemeng
    Wu, Jianlong
    Cheng, Yuan
    Nie, Liqiang
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 19405 - 19414
  • [44] Skew Class-Balanced Re-Weighting for Unbiased Scene Graph Generation
    Kang, Haeyong
    Yoo, Chang D.
    MACHINE LEARNING AND KNOWLEDGE EXTRACTION, 2023, 5 (01): : 287 - 303
  • [45] State-Aware Compositional Learning Toward Unbiased Training for Scene Graph Generation
    He, Tao
    Gao, Lianli
    Song, Jingkuan
    Li, Yuan-Fang
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 : 43 - 56
  • [46] SGT: Scene Graph-Guided Transformer for Surgical Report Generation
    Lin, Chen
    Zheng, Shuai
    Liu, Zhizhe
    Li, Youru
    Zhu, Zhenfeng
    Zhao, Yao
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2022, PT VII, 2022, 13437 : 507 - 518
  • [47] Tracklet Pair Proposal and Context Reasoning for Video Scene Graph Generation
    Jung, Gayoung
    Lee, Jonghun
    Kim, Incheol
    SENSORS, 2021, 21 (09)
  • [48] Meta Spatio-Temporal Debiasing for Video Scene Graph Generation
    Xu, Li
    Qu, Haoxuan
    Kuen, Jason
    Gu, Jiuxiang
    Liu, Jun
    COMPUTER VISION - ECCV 2022, PT XXVII, 2022, 13687 : 374 - 390
  • [49] DBiased-P: Dual-Biased Predicate Predictor for Unbiased Scene Graph Generation
    Han, Xianjing
    Song, Xuemeng
    Dong, Xingning
    Wei, Yinwei
    Liu, Meng
    Nie, Liqiang
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 5319 - 5329
  • [50] A Novel Two-Stage Training Method for Unbiased Scene Graph Generation via Distribution
    Jia, Dongdong
    Zhou, Meili
    Wei, Wei
    Wang, Dong
    Bai, Zongwen
    KSII TRANSACTIONS ON INTERNET AND INFORMATION SYSTEMS, 2023, 17 (12): : 3383 - 3397