Improved modeling of human vision by incorporating robustness to blur in convolutional neural networks

被引:1
|
作者
Jang, Hojin [1 ,2 ,3 ]
Tong, Frank [1 ]
机构
[1] Vanderbilt Univ, Vanderbilt Vis Res Ctr, Dept Psychol, Nashville, TN 37235 USA
[2] MIT, Dept Brain & Cognit Sci, Cambridge, MA 02193 USA
[3] Korea Univ, Dept Brain & Cognit Engn, Seoul, South Korea
基金
美国国家卫生研究院;
关键词
REPRESENTATIONS; RESPONSES; DYNAMICS; STIMULI; OBJECTS; CORTEX; AREA;
D O I
10.1038/s41467-024-45679-0
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Whenever a visual scene is cast onto the retina, much of it will appear degraded due to poor resolution in the periphery; moreover, optical defocus can cause blur in central vision. However, the pervasiveness of blurry or degraded input is typically overlooked in the training of convolutional neural networks (CNNs). We hypothesized that the absence of blurry training inputs may cause CNNs to rely excessively on high spatial frequency information for object recognition, thereby causing systematic deviations from biological vision. We evaluated this hypothesis by comparing standard CNNs with CNNs trained on a combination of clear and blurry images. We show that blur-trained CNNs outperform standard CNNs at predicting neural responses to objects across a variety of viewing conditions. Moreover, blur-trained CNNs acquire increased sensitivity to shape information and greater robustness to multiple forms of visual noise, leading to improved correspondence with human perception. Our results provide multi-faceted neurocomputational evidence that blurry visual experiences may be critical for conferring robustness to biological visual systems. The phenomenon of blurry or degraded visual input in humans has been overlooked in the training of convolutional neural networks (CNNs). Here, the authors show that blur-trained CNNs outperform standard CNNs in predicting neural responses to objects and show improved correspondence with human perception.
引用
收藏
页数:14
相关论文
共 50 条
  • [21] Adversarial Robustness of Multi-bit Convolutional Neural Networks
    Frickenstein, Lukas
    Sampath, Shambhavi Balamuthu
    Mori, Pierpaolo
    Vemparala, Manoj-Rohit
    Fasfous, Nael
    Frickenstein, Alexander
    Unger, Christian
    Passerone, Claudio
    Stechele, Walter
    [J]. INTELLIGENT SYSTEMS AND APPLICATIONS, VOL 3, INTELLISYS 2023, 2024, 824 : 157 - 174
  • [22] PROVABLE TRANSLATIONAL ROBUSTNESS FOR OBJECT DETECTION WITH CONVOLUTIONAL NEURAL NETWORKS
    Vierling, Axel
    James, Charu
    Berns, Karsten
    Katsaouni, Nikoletta
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 694 - 698
  • [23] Enhancing the robustness of the convolutional neural networks for traffic sign detection
    Khosravian, Amir
    Amirkhani, Abdollah
    Masih-Tehrani, Masoud
    [J]. PROCEEDINGS OF THE INSTITUTION OF MECHANICAL ENGINEERS PART D-JOURNAL OF AUTOMOBILE ENGINEERING, 2022, 236 (08) : 1849 - 1861
  • [24] IMPROVING THE ROBUSTNESS OF CONVOLUTIONAL NEURAL NETWORKS VIA SKETCH ATTENTION
    Chu, Tianshu
    Yang, Zuopeng
    Yang, Jie
    Huang, Xiaolin
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 869 - 873
  • [25] Estimation of motion blur kernel parameters using regression convolutional neural networks
    Varela, Luis G.
    Boucheron, Laura E.
    Sandoval, Steven
    Voelz, David
    Siddik, Abu Bucker
    [J]. JOURNAL OF ELECTRONIC IMAGING, 2024, 33 (02)
  • [26] On Vectorization of Deep Convolutional Neural Networks for Vision Tasks
    Ren, Jimmy S. J.
    Xu, Li
    [J]. PROCEEDINGS OF THE TWENTY-NINTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2015, : 1840 - 1846
  • [27] CMT: Convolutional Neural Networks Meet Vision Transformers
    Guo, Jianyuan
    Han, Kai
    Wu, Han
    Tang, Yehui
    Chen, Xinghao
    Wang, Yunhe
    Xu, Chang
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 12165 - 12175
  • [28] A improved pooling method for convolutional neural networks
    Zhao, Lei
    Zhang, Zhonglin
    [J]. SCIENTIFIC REPORTS, 2024, 14 (01)
  • [29] Visualization Comparison of Vision Transformers and Convolutional Neural Networks
    Shi, Rui
    Li, Tianxing
    Zhang, Liguo
    Yamaguchi, Yasushi
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 2327 - 2339
  • [30] A improved pooling method for convolutional neural networks
    Lei Zhao
    Zhonglin Zhang
    [J]. Scientific Reports, 14