Shuffle-Then-Assemble: Learning Object-Agnostic Visual Relationship Features

被引:29
|
作者
Yang, Xu [1 ]
Zhang, Hanwang [1 ]
Cai, Jianfei [1 ]
机构
[1] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore, Singapore
来源
关键词
D O I
10.1007/978-3-030-01258-8_3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Due to the fact that it is prohibitively expensive to completely annotate visual relationships, i.e., the (obj1, rel, obj2) triplets, relationship models are inevitably biased to object classes of limited pairwise patterns, leading to poor generalization to rare or unseen object combinations. Therefore, we are interested in learning object-agnostic visual features for more generalizable relationship models. By "agnostic", we mean that the feature is less likely biased to the classes of paired objects. To alleviate the bias, we propose a novel Shuffle-Then-Assemble pre-training strategy. First, we discard all the triplet relationship annotations in an image, leaving two unpaired object domains without obj1-obj2 alignment. Then, our feature learning is to recover possible obj1-obj2 pairs. In particular, we design a cycle of residual transformations between the two domains, to capture shared but not object-specific visual patterns. Extensive experiments on two visual relationship benchmarks show that by using our pre-trained features, naive relationship models can be consistently improved and even outperform other state-of-the-art relationship models. Code has been made available at: https://github.com/yangxuntu/vrd.
引用
收藏
页码:38 / 54
页数:17
相关论文
共 22 条
  • [1] Object-agnostic Affordance Categorization via Unsupervised Learning of Graph Embeddings
    Toumpa, Alexia
    Cohn, Anthony G.
    [J]. JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2023, 77 : 1 - 38
  • [2] Object-agnostic Affordance Categorization via Unsupervised Learning of Graph Embeddings
    Toumpa, Alexia
    Cohn, Anthony G.
    [J]. Journal of Artificial Intelligence Research, 2023, (77): : 1 - 38
  • [3] Object-Agnostic Vision Measurement Framework Based on One-Shot Learning and Behavior Tree
    Qin, Fangbo
    Xu, De
    Hannaford, Blake
    Hao, Tiantian
    [J]. IEEE TRANSACTIONS ON CYBERNETICS, 2023, 53 (08) : 5202 - 5215
  • [4] Contour Primitive of Interest Extraction Network Based on One-Shot Learning for Object-Agnostic Vision Measurement
    Qin, Fangbo
    Qin, Jie
    Huang, Siyu
    Xu, De
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 4311 - 4317
  • [5] CFVS: Coarse-to-Fine Visual Servoing for 6-DoF Object-Agnostic Peg-In-Hole Assembly
    Lu, Bo-Siang
    Chen, Tung-I
    Lee, Hsin-Ying
    Hsu, Winston H.
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2023), 2023, : 12402 - 12408
  • [6] Learning Relationship-Aware Visual Features
    Messina, Nicola
    Amato, Giuseppe
    Carrara, Fabio
    Falchi, Fabrizio
    Gennaro, Claudio
    [J]. COMPUTER VISION - ECCV 2018 WORKSHOPS, PT IV, 2019, 11132 : 486 - 501
  • [7] Object recognition by learning informative, biologically inspired visual features
    Wu, Yang
    Zheng, Nanning
    You, Qubo
    Du, Shaoyi
    [J]. 2007 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, VOLS 1-7, 2007, : 181 - 184
  • [8] Learning Visual Object Categories with Global Descriptors and Local Features
    Pereira, Rui
    Lopes, Luis Seabra
    [J]. PROGRESS IN ARTIFICIAL INTELLIGENCE, PROCEEDINGS, 2009, 5816 : 225 - 236
  • [9] Learning Visual Saliency Based on Object's Relative Relationship
    Wang, Senlin
    Zhao, Qi
    Song, Mingli
    Bu, Jiajun
    Chen, Chun
    Tao, Dacheng
    [J]. NEURAL INFORMATION PROCESSING, ICONIP 2012, PT V, 2012, 7667 : 318 - 327
  • [10] Learning deep domain-agnostic features from synthetic renders for industrial visual inspection
    Abubakr, Abdelrahman G.
    Jovancevic, Igor
    Mokhtari, Nour Islam
    Ben Abdallah, Hamdi
    Orteu, Jean-Jose
    [J]. JOURNAL OF ELECTRONIC IMAGING, 2022, 31 (05)