A developmental where-what network for concurrent and interactive visual attention and recognition

被引:3
|
作者
Ji, Zhengping [1 ]
Weng, Juyang [2 ]
机构
[1] Samsung Semicond Inc, Adv Image Res Lab ARIL, Pasadena, CA 91103 USA
[2] Michigan State Univ, Dept Comp Sci & Engn, E Lansing, MI 48824 USA
关键词
Developmental learning; Where-what sensorimotor pathways; Attention; Recognition; Brain-inspired neural network; MODEL; ALGORITHM; CORTEX; LAYERS;
D O I
10.1016/j.robot.2015.03.004
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper presents a brain-inspired developmental architecture called Where-What Network (WWN). In this second version of WWN, WWN-2 is learned for concurrent and interactive visual attention and recognition, via complementary pathways guided by "type" motor and "location" motor. The motor-driven top-down signals, together with bottom-up excitatory activities from the visual input, shape three possible information flows through a Y-shaped network. Using l(0) constrained sparse coding scheme, the top-down and bottom-up co-firing leads to a non-iterative cell-centered synaptic update model, entailing the strict entropy reduction from early to later layers, as well as a dual optimization of update directions and step sizes that dynamically depend on the firing ages of the neurons. Three operational modes for cluttered scenes emerge from the learning process, depending on what is available in the motor area: context-free mode for detection and recognition from a cluttered scene for a learned object, location-context mode for doing object recognition, and type-context mode for doing object search, all by a single network. To demonstrate the attention capabilities along with their interaction of visual processing, the proposed network is in the presence of complex backgrounds, learns on the fly, and produces engineering graded performance regarding attended pixel errors and recognition accuracy. As the proposed architecture is developmental, meaning that the internal representations are learned from pairs of input and motor signal, and thereby not manipulated internally for a specific task, we argue that the same learning principles and computational architecture can be potentially applicable to other sensory modalities, such as audition and touch. (C) 2015 Elsevier B.V. All rights reserved.
引用
收藏
页码:35 / 48
页数:14
相关论文
共 50 条
  • [41] An Improved Deep Neural Network Based on a Novel Visual Attention Mechanism for Text Recognition
    Nguyen Trong Thai
    Nguyen Hoang Thuan
    Dinh Viet Sang
    2021 RIVF INTERNATIONAL CONFERENCE ON COMPUTING AND COMMUNICATION TECHNOLOGIES (RIVF 2021), 2021, : 1 - 6
  • [42] DVAMN: Dual Visual Attention Matching Network for Zero-Shot Action Recognition
    Qi, Cheng
    Feng, Zhiyong
    Xing, Meng
    Su, Yong
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2021, PT V, 2021, 12895 : 564 - 575
  • [44] DACBN: Dual attention convolutional broad network for fine-grained visual recognition
    Chen, Tao
    Wang, Lijie
    Liu, Yang
    Yu, Haisheng
    PATTERN RECOGNITION, 2024, 156
  • [45] What is moving where? Infants' visual attention to dynamic objects may assist with processing of spatial relations
    Choi, Jihye
    Park, Youjeong
    FRONTIERS IN PSYCHOLOGY, 2024, 14
  • [46] Impaired recognition of interactive intentions in adults with autism spectrum disorder not attributable to differences in visual attention or coordination via eye contact and joint attention
    Jording, Mathis
    Hartz, Arne
    Vogel, David H. V.
    Schulte-Ruether, Martin
    Vogeley, Kai
    SCIENTIFIC REPORTS, 2024, 14 (01)
  • [47] The what-and-where filter - A spatial mapping neural network for object recognition and image understanding
    Carpenter, GA
    Grossberg, S
    Lesher, GW
    COMPUTER VISION AND IMAGE UNDERSTANDING, 1998, 69 (01) : 1 - 22
  • [48] Multi-scale fusion visual attention network for facial micro-expression recognition
    Pan, Hang
    Yang, Hongling
    Xie, Lun
    Wang, Zhiliang
    FRONTIERS IN NEUROSCIENCE, 2023, 17
  • [49] Zero-Shot Visual Recognition via Semantic Attention-Based Compare Network
    Nian, Fudong
    Sheng, Yikun
    Wang, Junfeng
    Li, Teng
    IEEE ACCESS, 2020, 8 : 26002 - 26011
  • [50] A brain-inspired object-based attention network for multiobject recognition and visual reasoning
    Adeli, Hossein
    Ahn, Seoyoung
    Zelinsky, Gregory J.
    JOURNAL OF VISION, 2023, 23 (05):