Gated Fusion of Discriminant Features for Caricature Recognition

被引:4
|
作者
Dai, Lingna [1 ]
Gao, Fei [1 ,2 ]
Li, Rongsheng [3 ]
Yu, Jiachen [1 ]
Shen, Xiaoyuan [1 ]
Xiong, Huilin [1 ]
Wu, Weilun [1 ]
机构
[1] Hangzhou Dianzi Univ, Sch Comp Sci & Technol, Key Lab Complex Syst Modeling & Simulat, Hangzhou 310018, Peoples R China
[2] Xidian Univ, Sch Elect Engn, State Key Lab Integrated Serv Networks, Xian 710071, Peoples R China
[3] State Grid Yantai Power Supply Co, Yantai 264001, Peoples R China
基金
中国国家自然科学基金;
关键词
Caricature; Face recognition; Deep learning; Convolutional block attention module; SphereFace;
D O I
10.1007/978-3-030-36189-1_47
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Caricature recognition is a challenging problem, because there are typically geometric deformations between photographs and caricatures. It is nontrivial to learn discriminant large-margin features. To combat this challenge, we propose a novel framework by using a gated fusion of global and local discriminant features. First, we employ A-Softmax loss to jointly learn angularly discriminant features of the whole face and local facial parts. Besides, we use the convolutional block attention module (CBAM) to further boost the discriminant ability of the learnt features. Next, we use global features as dominant representation and local features as supplemental ones; and propose a gated fusion unit to automatically learn the weighting factors for these local parts and moderate local features correspondingly. Finally, an integration of all these features is used for caricature recognition. Extensive experiments are conducted on the cross-modal face recognition task. Results show that, our method significantly boosts previous state-of-the-art Rank-1 and Rank-10 from 36.27% to 55.29% and from 64.37% to 85.78%, respectively, for caricature-to-photograph (C2P) recognition. Besides, our method achieves a Rank-1 of 60.81% and Rank-10 of 89.26% for photograph-to-caricature (P2C) recognition.
引用
收藏
页码:563 / 573
页数:11
相关论文
共 50 条
  • [1] Fusion of Directional Spatial Discriminant Features for Face Recognition
    Dey, Aniruddha
    Sing, Jamuna Kanta
    Chowdhury, Shiladitya
    [J]. FIRST INTERNATIONAL CONFERENCE ON COMPUTATIONAL INTELLIGENCE: MODELING TECHNIQUES AND APPLICATIONS (CIMTA) 2013, 2013, 10 : 747 - 754
  • [2] Independent Gabor Analysis of Discriminant Features Fusion for Face Recognition
    Wu, Jiying
    An, Gaoyun
    Ruan, Qiuqi
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2009, 16 (1-3) : 97 - 100
  • [3] Face Recognition by Cognitive Discriminant Features
    Firouzian, Iman
    Firouzian, Nematallah
    [J]. INTERNATIONAL JOURNAL OF NONLINEAR ANALYSIS AND APPLICATIONS, 2020, 11 (01): : 7 - 20
  • [4] Discriminant Spectrotemporal Features for Phoneme Recognition
    Mesgarani, Nima
    Sivaram, G. S. V. S.
    Nemala, Sridhar Krishna
    Elhilali, Mounya
    Hermansky, Hynek
    [J]. INTERSPEECH 2009: 10TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2009, VOLS 1-5, 2009, : 2947 - 2950
  • [5] Using kernel discriminant analysis and 2DGabor local features fusion for face recognition
    Lin, KeZheng
    Xu, Ying
    Yuan, Zhong
    [J]. International Journal of Digital Content Technology and its Applications, 2010, 4 (08) : 232 - 241
  • [6] Face recognition by inverse Fisher discriminant features
    Zhuang, XS
    Dai, DQ
    Yuen, PC
    [J]. ADVANCES IN BIOMETRICS, PROCEEDINGS, 2006, 3832 : 92 - 98
  • [7] CARICATURE AND FACE RECOGNITION
    MAURO, R
    KUBOVY, M
    [J]. MEMORY & COGNITION, 1992, 20 (04) : 433 - 440
  • [8] Caricature, recognition, misrepresentation
    Fantelli, Federico
    [J]. PHENOMENOLOGY AND THE COGNITIVE SCIENCES, 2024,
  • [9] A RELATION NETWORK EMBEDDED WITH PRIOR FEATURES FOR FEW-SHOT CARICATURE RECOGNITION
    Zheng, Wenbo
    Yan, Lan
    Gou, Chao
    Zhang, Wenwen
    Wang, Fei-Yue
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2019, : 1510 - 1515
  • [10] Caricature recognition in a neural network
    Tanaka, JW
    Simon, VB
    [J]. VISUAL COGNITION, 1996, 3 (04) : 305 - 324