FairScene: Learning unbiased object interactions for indoor scene synthesis

被引:0
|
作者
Wu, Zhenyu [1 ]
Wang, Ziwei [2 ]
Liu, Shengyu [2 ]
Luo, Hao [1 ]
Lu, Jiwen [2 ]
Yan, Haibin [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Sch Intelligent Engn & Automat, Beijing 100876, Peoples R China
[2] Tsinghua Univ, Dept Automat, Beijing 100084, Peoples R China
基金
中国国家自然科学基金;
关键词
Indoor scene synthesis; Graph neural networks; Causal inference;
D O I
10.1016/j.patcog.2024.110737
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we propose an unbiased graph neural network learning method called FairScene for indoor scene synthesis. Conventional methods directly apply graphical models to represent the correlation of objects for subsequent furniture insertion. However, due to the object category imbalance in dataset collection and complex object entanglement with implicit confounders, these methods usually generate significantly biased scenes. Moreover, the performance of these methods varies greatly for different indoor scenes. To address this, we propose a framework named FairScene which can fully exploit unbiased object interactions through causal reasoning, so that fair scene synthesis is achieved by calibrating the long-tailed category distribution and mitigating the confounder effects. Specifically, we remove the long-tailed object priors subtract the counterfactual prediction obtained from default input, and intervene in the input feature by cutting off the causal link to confounders based on the causal graph. Extensive experiments on the 3D-FRONT dataset show that our proposed method outperforms the state-of-the-art indoor scene generation methods and enhances vanilla models on a wide variety of vision tasks including scene completion and object recognition.
引用
收藏
页数:13
相关论文
共 50 条
  • [41] Indoor Scene Object Detection Based on Improved YOLOv4 Algorithm
    Li Weigang
    Yang Chao
    Jiang Lin
    Zhao Yuntao
    LASER & OPTOELECTRONICS PROGRESS, 2022, 59 (18)
  • [42] Learning Object-to-Class Kernels for Scene Classification
    Zhang, Lei
    Zhen, Xiantong
    Shao, Ling
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2014, 23 (08) : 3241 - 3253
  • [43] PCPL: Predicate-Correlation Perception Learning for Unbiased Scene Graph Generation
    Yan, Shaotian
    Shen, Chen
    Jin, Zhongming
    Huang, Jianqiang
    Jiang, Rongxin
    Chen, Yaowu
    Hua, Xian-Sheng
    MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 265 - 273
  • [44] Learning depth-aware features for indoor scene understanding
    Suting Chen
    Dongwei Shao
    Liangchen Zhang
    Chuang Zhang
    Multimedia Tools and Applications, 2022, 81 : 42573 - 42590
  • [45] Semantic embedding for indoor scene recognition by weighted hypergraph learning
    Yu, Jun
    Hong, Chaoqun
    Tao, Dapeng
    Wang, Meng
    SIGNAL PROCESSING, 2015, 112 : 129 - 136
  • [46] Learning Informative Edge Maps for Indoor Scene Layout Prediction
    Mallya, Arun
    Lazebnik, Svetlana
    2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, : 936 - 944
  • [47] A Deep Learning based Scene Recognition Algorithm for Indoor Localization
    Labinghisa, Boney
    Lee, Dong Myung
    3RD INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE IN INFORMATION AND COMMUNICATION (IEEE ICAIIC 2021), 2021, : 167 - 170
  • [48] Learning depth-aware features for indoor scene understanding
    Chen, Suting
    Shao, Dongwei
    Zhang, Liangchen
    Zhang, Chuang
    MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (29) : 42573 - 42590
  • [49] Deep Learning Framework for Scene Based Indoor Location Recognition
    Hanni, Akkamahadevi
    Chickerur, Satyadhyan
    Bidari, Indira
    PROCEEDINGS OF 2017 INTERNATIONAL CONFERENCE ON TECHNOLOGICAL ADVANCEMENTS IN POWER AND ENERGY (TAP ENERGY): EXPLORING ENERGY SOLUTIONS FOR AN INTELLIGENT POWER GRID, 2017,
  • [50] Discriminative Learning with Latent Variables for Cluttered Indoor Scene Understanding
    Wang, Huayan
    Gould, Stephen
    Koller, Daphne
    COMPUTER VISION-ECCV 2010, PT II, 2010, 6312 : 435 - +