Top-Down Visual Saliency via Joint CRF and Dictionary Learning

被引:98
|
作者
Yang, Jimei [1 ]
Yang, Ming-Hsuan [2 ]
机构
[1] Adobe Res, San Jose, CA 95110 USA
[2] Univ Calif Merced, Sch Engn, Merced, CA USA
基金
美国国家科学基金会;
关键词
Visual saliency; top-down visual saliency; fixation prediction; dictionary learning and conditional random fields; FEATURES; ATTENTION;
D O I
10.1109/TPAMI.2016.2547384
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Top-down visual saliency is an important module of visual attention. In this work, we propose a novel top-down saliency model that jointly learns a Conditional Random Field (CRF) and a visual dictionary. The proposed model incorporates a layered structure from top to bottom: CRF, sparse coding and image patches. With sparse coding as an intermediate layer, CRF is learned in a feature-adaptive manner; meanwhile with CRF as the output layer, the dictionary is learned under structured supervision. For efficient and effective joint learning, we develop a max-margin approach via a stochastic gradient descent algorithm. Experimental results on the Graz-02 and PASCAL VOC datasets show that our model performs favorably against state-of-the-art top-down saliency methods for target object localization. In addition, the dictionary update significantly improves the performance of our model. We demonstrate the merits of the proposed top-down saliency model by applying it to prioritizing object proposals for detection and predicting human fixations.
引用
收藏
页码:576 / 588
页数:13
相关论文
共 50 条
  • [31] Bottom-up saliency and top-down attention interact inhuman primary visual cortex
    Hopf, JM
    Noesselt, T
    Tempelmann, C
    Kai, B
    Hans-Jochen, H
    Jochen, B
    JOURNAL OF COGNITIVE NEUROSCIENCE, 2002, : 97 - 98
  • [32] Integrating bottom-up and top-down visual stimulus for saliency detection in news video
    Wu, Bo
    Xu, Linfeng
    MULTIMEDIA TOOLS AND APPLICATIONS, 2014, 73 (03) : 1053 - 1075
  • [33] Top-down facilitation of visual recognition
    Bar, M
    Kassam, KS
    Ghuman, AS
    Boshyan, J
    Schmidt, AM
    Dale, AM
    Hämäläinen, MS
    Marinkovic, K
    Schacter, DL
    Rosen, BR
    Halgren, E
    PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2006, 103 (02) : 449 - 454
  • [34] Top-down control of visual attention
    Noudoost, Behrad
    Chang, Mindy H.
    Steinmetz, Nicholas A.
    Moore, Tirin
    CURRENT OPINION IN NEUROBIOLOGY, 2010, 20 (02) : 183 - 190
  • [35] Perceptual learning via modification of cortical top-down signals
    Schaefer, Roland
    Vasilaki, Eleni
    Senn, Walter
    PLOS COMPUTATIONAL BIOLOGY, 2007, 3 (08) : 1555 - 1566
  • [36] Learning enhances the relative impact of top-down processing in the visual cortex
    Makino, Hiroshi
    Komiyama, Takaki
    NATURE NEUROSCIENCE, 2015, 18 (08) : 1116 - +
  • [37] CrowdTrans: Learning top-down visual perception for crowd counting by transformer
    Guo, Weiyu
    Yang, Shaopeng
    Ren, Yuheng
    Huang, Yongzhen
    NEUROCOMPUTING, 2024, 587
  • [38] Top-down attentional guidance based on implicit learning of visual covariation
    Chun, MM
    Jiang, YH
    PSYCHOLOGICAL SCIENCE, 1999, 10 (04) : 360 - 365
  • [39] Layout-Driven Top-Down Saliency Detection for Webpage
    Li, Xixi
    Liu, Di
    Zhang, Kao
    Chen, Zhenzhong
    ADVANCES IN MULTIMEDIA INFORMATION PROCESSING - PCM 2017, PT II, 2018, 10736 : 438 - 446
  • [40] Top-down based saliency model in traffic driving environment
    Deng, Tao
    Chen, Andong
    Gao, Min
    Yan, Hongmei
    2014 IEEE 17TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2014, : 75 - 80