Frame-GAN: Increasing the frame rate of gait videos with generative adversarial networks

被引:9
|
作者
Xue, Wei [1 ,2 ]
Ai, Hong [1 ]
Sun, Tianyu [2 ]
Song, Chunfeng [2 ,3 ]
Huang, Yan [2 ,3 ]
Wang, Liang [2 ,3 ]
机构
[1] Harbin Univ Sci & Technol, Sch Automat, Harbin 150001, Peoples R China
[2] Chinese Acad Sci CASIA, CRIPAC, Inst Automat, Natl Lab Pattern Recognit NLPR, Beijing 100190, Peoples R China
[3] UCAS, Beijing 100190, Peoples R China
基金
中国国家自然科学基金;
关键词
Gait recognition; Generative adversarial networks; Metric learning; Deep learning; RECOGNITION; IMAGE;
D O I
10.1016/j.neucom.2019.11.015
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Most existing methods of identifying person except gait recognition require the cooperation of the subjects. Aiming at detecting the pattern of human walking movement, gait recognition takes advantage of the time-serial data and can identify a person distantly. The time-serial data, which is usually presented in video form, always has a limitation in frame rate, which intrinsically affects the performance of the recognition models. In order to increase the frame rate of gait videos, we propose a new kind of generative adversarial networks (GAN) named Frame-GAN to reduce the gap between adjacent frames. Inspired by the recent advances in metric learning, we also propose a new effective loss function named Margin Ratio Loss (MRL) to boost the recognition model. We evaluate the proposed method on the challenging CASIA-B and OU-ISIR gait databases. Extensive experimental results show that the proposed Frame-GAN and MRL are effective. (C) 2019 Elsevier B.V. All rights reserved.
引用
收藏
页码:95 / 104
页数:10
相关论文
共 50 条
  • [11] Increasing Detection Rate for Imbalanced Malicious Traffic using Generative Adversarial Networks
    Memmesheimer, Pascal
    Machmeier, Stefan
    Heuveline, Vincent
    PROCEEDINGS OF THE 2024 EUROPEAN INTERDISCIPLINARY CYBERSECURITY CONFERENCE, EICC 2024, 2024, : 74 - 81
  • [12] Inter-frame video image generation based on spatial continuity generative adversarial networks
    Zhang, Tao
    Jiang, Peipei
    Zhang, Meng
    SIGNAL IMAGE AND VIDEO PROCESSING, 2019, 13 (08) : 1487 - 1494
  • [13] Inter-frame video image generation based on spatial continuity generative adversarial networks
    Tao Zhang
    Peipei Jiang
    Meng Zhang
    Signal, Image and Video Processing, 2019, 13 : 1487 - 1494
  • [14] Frame Rate Up-Conversion in Echocardiography Using a Conditioned Variational Autoencoder and Generative Adversarial Model
    Dezaki, Fatemeh Taheri
    Girgis, Hany
    Rohling, Robert
    Gin, Ken
    Abolmaesumi, Purang
    Tsang, Teresa
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2019, PT II, 2019, 11765 : 705 - 713
  • [15] Traffic Analytics With Low-Frame-Rate Videos
    Luo, Zhiming
    Jodoin, Pierre-Marc
    Su, Song-Zhi
    Li, Shao-Zi
    Larochelle, Hugo
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2018, 28 (04) : 878 - 891
  • [16] Dual generative adversarial networks for automated component layout design of steel frame-brace structures
    Fu, Bochao
    Gao, Yuqing
    Wang, Wei
    AUTOMATION IN CONSTRUCTION, 2023, 146
  • [17] Generative adversarial networks for generating RGB-D videos
    Nakahira, Yuki
    Kawamoto, Kazuhiko
    2018 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2018, : 1276 - 1281
  • [18] FRAME ATTENTION NETWORKS FOR FACIAL EXPRESSION RECOGNITION IN VIDEOS
    Meng, Debin
    Peng, Xiaojiang
    Wang, Kai
    Qiao, Yu
    2019 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2019, : 3866 - 3870
  • [19] Generative Adversarial Network-Based Frame Extrapolation for Video Coding
    Lin, Jianping
    Liu, Dong
    Li, Houqiang
    Wu, Feng
    2018 IEEE INTERNATIONAL CONFERENCE ON VISUAL COMMUNICATIONS AND IMAGE PROCESSING (IEEE VCIP), 2018,
  • [20] Generative Adversarial Networks for Increasing the Veracity of Big Data
    Dering, Matthew L.
    Tucker, Conrad S.
    2017 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2017, : 2595 - 2602