Training Deep Networks from Zero to Hero: avoiding pitfalls and going beyond

被引:7
|
作者
Ponti, Moacir A. [1 ]
dos Santos, Fernando P. [1 ]
Ribeiro, Leo S. F. [1 ]
Cavallari, Gabriel B. [1 ]
机构
[1] ICMC Univ Sao Paulo USP, Sao Carlos, SP, Brazil
基金
巴西圣保罗研究基金会;
关键词
D O I
10.1109/SIBGRAPI54419.2021.00011
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Training deep neural networks may be challenging in real world data. Using models as black-boxes, even with transfer learning, can result in poor generalization or inconclusive results when it comes to small datasets or specific applications. This tutorial covers the basic steps as well as more recent options to improve models, in particular, but not restricted to, supervised learning. It can be particularly useful in datasets that are not as well-prepared as those in challenges, and also under scarce annotation and/or small data. We describe basic procedures as data preparation, optimization and transfer learning, but also recent architectural choices such as use of transformer modules, alternative convolutional layers, activation functions, wide/depth, as well as training procedures including curriculum, contrastive and self-supervised learning.
引用
收藏
页码:9 / 16
页数:8
相关论文
共 23 条
  • [1] Going Beyond Saliency Maps: Training Deep Models to Interpret Deep Models
    Liu, Zixuan
    Adeli, Ehsan
    Pohl, Kilian M.
    Zhao, Qingyu
    INFORMATION PROCESSING IN MEDICAL IMAGING, IPMI 2021, 2021, 12729 : 71 - 82
  • [2] Beyond the golden hour - Avoiding the pitfalls from resuscitation to critical care
    Schinco, M
    Tepas, JJ
    SURGICAL CLINICS OF NORTH AMERICA, 2002, 82 (02) : 325 - +
  • [3] Going Deep: The Role of Neural Networks for Renal Survival and Beyond
    Averitt, Amelia J.
    Natarajan, Karthik
    KIDNEY INTERNATIONAL REPORTS, 2018, 3 (02): : 242 - 243
  • [4] From Zero to Hero: Generating Training Data for Question-To-Cypher Models
    Opitz, Dominik
    Hochgeschwender, Nico
    2022 IEEE/ACM 1ST INTERNATIONAL WORKSHOP ON NATURAL LANGUAGE-BASED SOFTWARE ENGINEERING (NLBSE 2022), 2022, : 17 - 20
  • [5] Eigendecomposition-Free Training of Deep Networks with Zero Eigenvalue-Based Losses
    Dang, Zheng
    Yi, Kwang Moo
    Hu, Yinlin
    Wang, Fei
    Fua, Pascal
    Salzmann, Mathieu
    COMPUTER VISION - ECCV 2018, PT V, 2018, 11209 : 792 - 807
  • [6] From Pioneering Artificial Neural Networks to Deep Learning and Beyond
    Foresti, Gian Luca
    INTERNATIONAL JOURNAL OF NEURAL SYSTEMS, 2021, 31 (05)
  • [7] Structured and Systematic Team and Procedure Training in Severe Trauma: Going from 'Zero to Hero' for a Time-Critical, Low-Volume Emergency Procedure Over Three Time Periods
    Meshkinfamfard, Maryam
    Narvestad, Jon Kristian
    Larsen, Johannes Wiik
    Kanani, Arezo
    Vennesland, Jorgen
    Reite, Andreas
    Vetrhus, Morten
    Thorsen, Kenneth
    Soreide, Kjetil
    WORLD JOURNAL OF SURGERY, 2021, 45 (05) : 1340 - 1348
  • [8] Structured and Systematic Team and Procedure Training in Severe Trauma: Going from ‘Zero to Hero’ for a Time-Critical, Low-Volume Emergency Procedure Over Three Time Periods
    Maryam Meshkinfamfard
    Jon Kristian Narvestad
    Johannes Wiik Larsen
    Arezo Kanani
    Jørgen Vennesland
    Andreas Reite
    Morten Vetrhus
    Kenneth Thorsen
    Kjetil Søreide
    World Journal of Surgery, 2021, 45 : 1340 - 1348
  • [9] Global Optimality Beyond Two Layers: Training Deep ReLU Networks via Convex Programs
    Ergen, Tolga
    Pilanci, Mert
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [10] Training Spiking Neural Networks Using Lessons From Deep Learning
    Eshraghian, Jason K.
    Ward, Max
    Neftci, Emre O.
    Wang, Xinxin
    Lenz, Gregor
    Dwivedi, Girish
    Bennamoun, Mohammed
    Jeong, Doo Seok
    Lu, Wei D.
    PROCEEDINGS OF THE IEEE, 2023, 111 (09) : 1016 - 1054