Situational Perception Guided Image Matting

被引:0
|
作者
Xu, Bo [1 ]
Xie, Jiake [2 ]
Huang, Han [1 ]
Li, Ziwen [1 ]
Lu, Cheng [3 ]
Tang, Yong [2 ]
Guo, Yandong [1 ]
机构
[1] OPPO Res Inst, Beijing, Peoples R China
[2] PicUp AI, Hangzhou, Zhejiang, Peoples R China
[3] Xpeng Motors, Cupertino, CA USA
关键词
image matting; trimap; visual-to-textual; cross modality; transformer;
D O I
10.1145/3503161.3548036
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Most automatic matting methods try to separate the salient foreground from the background. However, the insufficient quantity and subjective bias of the current existing matting datasets make it difficult to fully explore the semantic association between object-to-object and object-to-environment in a given image. In this paper, we propose a Situational Perception Guided Image Matting (SPG-IM) method that mitigates subjective bias of matting annotations and captures sufficient situational perception information for better global saliency distilled from the visual-to-textual task. SPG-IM can better associate inter-objects and object-to-environment saliency, and compensate the subjective nature of image matting and its expensive annotation. We also introduce a textual Semantic Transformation (TST) module that can effectively transform and integrate the semantic feature stream to guide the visual representations. In addition, an Adaptive Focal Transformation (AFT) Refinement Network is proposed to adaptively switch multi-scale receptive fields and focal points to enhance both global and local details. Extensive experiments demonstrate the effectiveness of situational perception guidance from the visual-to-textual tasks on image matting, and our model outperforms the state-of-the-art methods. We also analyze the significance of different components in our model.
引用
收藏
页码:5283 / 5293
页数:11
相关论文
共 50 条
  • [1] Flexible Interactive Guided Image Matting
    Cheng, Hang
    Xu, Shugong
    Guo, Fengjun
    IEEE ACCESS, 2023, 11 : 58808 - 58821
  • [2] Image matting in the perception granular deep learning
    Hu, Hong
    Pang, Liang
    Shi, Zhongzhi
    KNOWLEDGE-BASED SYSTEMS, 2016, 102 : 51 - 63
  • [3] Text-Guided Portrait Image Matting
    Xu Y.
    Yao X.
    Liu B.
    Quan Y.
    Ji H.
    IEEE Transactions on Artificial Intelligence, 2024, 5 (08): : 1 - 13
  • [4] Perception based alpha estimation in natural image matting
    Lin, SY
    Shi, RY
    2004 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXP (ICME), VOLS 1-3, 2004, : 1407 - 1410
  • [5] Natural Image Matting via Guided Contextual Attention
    Li, Yaoyi
    Lu, Hongtao
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 11450 - 11457
  • [6] AlphaNet: An Attention Guided Deep Network for Automatic Image Matting
    Sharma, Rishab
    Deora, Rahul
    Vishvakarma, Anirudha
    2020 INTERNATIONAL CONFERENCE ON OMNI-LAYER INTELLIGENT SYSTEMS (IEEE COINS 2020), 2020, : 174 - 181
  • [7] Mutli-focus image fusion based on guided filter and image matting network
    Zhu P.
    Li X.
    Wang P.
    Jin X.
    Yao S.
    Multimedia Tools and Applications, 2025, 84 (3) : 1239 - 1259
  • [8] Multi-guided-based image matting via boundary detection
    Yao, Guilin
    Sun, Anming
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2024, 243
  • [9] Improved Minimum Spanning Tree based Image Segmentation with Guided Matting
    Wang, Weixing
    Tu, Angyan
    Bergholm, Fredrik
    KSII TRANSACTIONS ON INTERNET AND INFORMATION SYSTEMS, 2022, 16 (01): : 211 - 230
  • [10] Wider and Higher: Intensive Integration and Global Foreground Perception for Image Matting
    Qiao, Yu
    Wei, Ziqi
    Liu, Yuhao
    Wang, Yuxin
    Zhou, Dongsheng
    Zhang, Qiang
    Yang, Xin
    ADVANCES IN COMPUTER GRAPHICS, CGI 2022, 2022, 13443 : 541 - 553