Modout: Learning Multi-modal Architectures by Stochastic Regularization

被引:9
|
作者
Li, Fan [1 ]
Neverova, Natalia [2 ]
Wolf, Christian [3 ]
Taylor, Graham [1 ]
机构
[1] Univ Guelph, Sch Engn, Guelph, ON, Canada
[2] Facebook, Paris, France
[3] INSA Lyon, LIRIS, Lyon, France
关键词
D O I
10.1109/FG.2017.59
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Model selection methods based on stochastic regularization have been widely used in deep learning due to their simplicity and effectiveness. The well-known Dropout method treats all units, visible or hidden, in the same way, thus ignoring any a priori information related to grouping or structure. Such structure is present in multi-modal learning applications such as affect analysis and gesture recognition, where subsets of units may correspond to individual modalities. Here we describe Modout, a model selection method based on stochastic regularization, which is particularly useful in the multi-modal setting. Different from other forms of stochastic regularization, it is capable of learning whether or when to fuse two modalities in a layer, which is usually considered to be an architectural hyper-parameter by deep learning researchers and practitioners. Modout is evaluated on two real multi-modal datasets. The results indicate improved performance compared to other forms of stochastic regularization. The result on the Montalbano dataset shows that learning a fusion structure by Modout is on par with a state-of-the-art carefully designed architecture.
引用
收藏
页码:422 / 429
页数:8
相关论文
共 50 条
  • [11] Multi-modal and multi-granular learning
    Zhang, Bo
    Zhang, Ling
    [J]. ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PROCEEDINGS, 2007, 4426 : 9 - +
  • [12] Learning in an Inclusive Multi-Modal Environment
    Graham, Deryn
    Benest, Ian
    Nicholl, Peter
    [J]. JOURNAL OF CASES ON INFORMATION TECHNOLOGY, 2010, 12 (03) : 28 - 44
  • [13] Learning of Multi-Modal Stimuli in Hawkmoths
    Balkenius, Anna
    Dacke, Marie
    [J]. PLOS ONE, 2013, 8 (07):
  • [14] Reliable Multi-modal Learning: A Survey
    Yang, Yang
    Zhan, De-Chuan
    Jiang, Yuan
    Xiong, Hui
    [J]. Ruan Jian Xue Bao/Journal of Software, 2021, 32 (04): : 1067 - 1081
  • [15] Multi-Modal Meta Continual Learning
    Gai, Sibo
    Chen, Zhengyu
    Wang, Donglin
    [J]. 2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [16] Imagery in multi-modal object learning
    Jüttner, M
    Rentschler, I
    [J]. BEHAVIORAL AND BRAIN SCIENCES, 2002, 25 (02) : 197 - +
  • [17] Learning multi-modal control programs
    Mehta, TR
    Egerstedt, M
    [J]. HYBRID SYSTEMS: COMPUTATION AND CONTROL, 2005, 3414 : 466 - 479
  • [18] MULTI-MODAL LEARNING FOR GESTURE RECOGNITION
    Cao, Congqi
    Zhang, Yifan
    Lu, Hanqing
    [J]. 2015 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO (ICME), 2015,
  • [19] Multi-modal Network Representation Learning
    Zhang, Chuxu
    Jiang, Meng
    Zhang, Xiangliang
    Ye, Yanfang
    Chawla, Nitesh, V
    [J]. KDD '20: PROCEEDINGS OF THE 26TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2020, : 3557 - 3558
  • [20] Modelling multi-modal learning in a hawkmoth
    Balkenius, Anna
    Kelber, Almut
    Balkenius, Christian
    [J]. FROM ANIMALS TO ANIMATS 9, PROCEEDINGS, 2006, 4095 : 422 - 433