Transformers Meet Small Datasets

被引:5
|
作者
Shao, Ran [1 ,2 ]
Bi, Xiao-Jun [3 ]
机构
[1] Harbin Engn Univ, Coll Informat & Commun Engn, Harbin 150001, Peoples R China
[2] Harbin Vocat & Tech Coll, Coll Elect & Informat Engn, Harbin 150001, Peoples R China
[3] Minzu Univ China, Dept Informat Engn, Beijing 100081, Peoples R China
关键词
Convolutional neural networks; small datasets; transformer; vision transformer;
D O I
10.1109/ACCESS.2022.3221138
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The research and application areas of transformers have been extensively enlarged due to the success of vision transformers (ViTs). However, due to the lack of local content acquisition capabilities, the pure transformer architectures cannot be trained directly on small datasets. In this work, we first propose a new hybrid model by combining the transformer and convolution neural network (CNN). The proposed model improves the classification ability on small datasets. This is accomplished by introducing more convolution operations in the transformer's two core sections: 1) Instead of the original multi-head attention mechanism, we design a convolutional parameter sharing multi-head attention (CPSA) block that incorporates the convolutional parameter sharing projection in the attention mechanism; 2) the feed-forward network in each transformer encoder block is replaced with a local feed-forward network (LFFN) block that introduces a sandglass block with more depth-wise convolutions to provide more locality to the transformers. We achieve state-of-the-art results when training from scratch on 4 small datasets as compared with the transformers and CNNs without extensive computing resources and auxiliary training. The proposed strategy opens up new paths for the application of transformers on small datasets.
引用
收藏
页码:118454 / 118464
页数:11
相关论文
共 50 条
  • [1] Optimizing Deeper Transformers on Small Datasets
    Xu, Peng
    Kumar, Dhruv
    Yang, Wei
    Zi, Wenjie
    Tang, Keyi
    Huang, Chenyang
    Cheung, Jackie Chi Kit
    Prince, Simon J. D.
    Cao, Yanshuai
    [J]. 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1 (ACL-IJCNLP 2021), 2021, : 2089 - 2102
  • [2] Efficient Training of Visual Transformers with Small Datasets
    Liu, Yahui
    Sangineto, Enver
    Bi, Wei
    Sebe, Nicu
    Lepri, Bruno
    De Nadai, Marco
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [3] Accumulated Trivial Attention Matters in Vision Transformers on Small Datasets
    Chen, Xiangyu
    Hu, Qinghao
    Li, Kaidong
    Zhong, Cuncong
    Wang, Guanghui
    [J]. 2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 3973 - 3981
  • [4] AViT: Adapting Vision Transformers for Small Skin Lesion Segmentation Datasets
    Du, Siyi
    Bayasi, Nourhan
    Hamarneh, Ghassan
    Garbi, Rafeef
    [J]. MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2023 WORKSHOPS, 2023, 14393 : 25 - 36
  • [5] Exploring Advances in Transformers and CNN for Skin Lesion Diagnosis on Small Datasets
    de Lima, Leandro M.
    Krohling, Renato A.
    [J]. INTELLIGENT SYSTEMS, PT II, 2022, 13654 : 282 - 296
  • [6] Vision Transformers for Small Histological Datasets Learned Through Knowledge Distillation
    Kanwal, Neel
    Eftestol, Trygve
    Khoraminia, Farbod
    Zuiverloon, Tahlita C. M.
    Engan, Kjersti
    [J]. ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2023, PT III, 2023, 13937 : 167 - 179
  • [7] Bridging the Gap Between Vision Transformers and Convolutional Neural Networks on Small Datasets
    Lu, Zhiying
    Xie, Hongtao
    Liu, Chuanbin
    Zhang, Yongdong
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [8] Fine-Grained Fish Classification From Small to Large Datasets With Vision Transformers
    Veiga, Ricardo J. M.
    Rodrigues, Joao M. F.
    [J]. IEEE ACCESS, 2024, 12 : 113642 - 113660
  • [9] Enhancing performance of vision transformers on small datasets through local inductive bias incorporation
    Akkaya I.B.
    Kathiresan S.S.
    Arani E.
    Zonooz B.
    [J]. Pattern Recognition, 2024, 153
  • [10] Quasi Character-Level Transformers to Improve Neural Machine Translation on Small Datasets
    Carrion, Salvador
    Casacuberta, Francisco
    [J]. 2021 EIGHTH INTERNATIONAL CONFERENCE ON SOCIAL NETWORK ANALYSIS, MANAGEMENT AND SECURITY (SNAMS), 2021, : 184 - 189