In-domain versus out-of-domain transfer learning in plankton image classification

被引:0
|
作者
Andrea Maracani
Vito Paolo Pastore
Lorenzo Natale
Lorenzo Rosasco
Francesca Odone
机构
[1] Istituto Italiano di Tecnologia,
[2] MaLGa-DIBRIS,undefined
[3] Università degli studi di Genova,undefined
[4] CBMM,undefined
[5] Massachusetts Institute of Technology,undefined
来源
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Plankton microorganisms play a huge role in the aquatic food web. Recently, it has been proposed to use plankton as a biosensor, since they can react to even minimal perturbations of the aquatic environment with specific physiological changes, which may lead to alterations in morphology and behavior. Nowadays, the development of high-resolution in-situ automatic acquisition systems allows the research community to obtain a large amount of plankton image data. Fundamental examples are the ZooScan and Woods Hole Oceanographic Institution (WHOI) datasets, comprising up to millions of plankton images. However, obtaining unbiased annotations is expensive both in terms of time and resources, and in-situ acquired datasets generally suffer from severe imbalance, with only a few images available for several species. Transfer learning is a popular solution to these challenges, with ImageNet1K being the most-used source dataset for pre-training. On the other hand, datasets like the ZooScan and the WHOI may represent a valuable opportunity to compare out-of-domain and large-scale plankton in-domain source datasets, in terms of performance for the task at hand.In this paper, we design three transfer learning pipelines for plankton image classification, with the aim of comparing in-domain and out-of-domain transfer learning on three popular benchmark plankton datasets. The general framework consists in fine-tuning a pre-trained model on a plankton target dataset. In the first pipeline, the model is pre-trained from scratch on a large-scale plankton dataset, in the second, it is pre-trained on large-scale natural image datasets (ImageNet1K or ImageNet22K), while in the third, a two-stage fine-tuning is implemented (ImageNet →\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\rightarrow $$\end{document} large-scale plankton dataset →\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\rightarrow $$\end{document} target plankton dataset). Our results show that an out-of-domain ImageNet22K pre-training outperforms the plankton in-domain ones, with an average boost in test accuracy of around 6%. In the next part of this work, we adopt three ImageNet22k pre-trained Vision Transformers and one ConvNeXt, obtaining results on par (or slightly superior) with the state-of-the-art, corresponding to the usage of CNN models ensembles, with a single model. Finally, we design and test an ensemble of our Vision Transformers and the ConvNeXt, outperforming the state-of-the-art existing works on plankton image classification on the three target datasets. To support scientific community contribution and further research, our implemented code is open-source and available at https://github.com/Malga-Vision/plankton_transfer.
引用
收藏
相关论文
共 50 条
  • [1] In-domain versus out-of-domain transfer learning in plankton image classification
    Maracani, Andrea
    Pastore, Vito Paolo
    Natale, Lorenzo
    Rosasco, Lorenzo
    Odone, Francesca
    [J]. SCIENTIFIC REPORTS, 2023, 13 (01)
  • [2] In-domain versus out-of-domain transfer learning for document layout analysis
    De Nardin, Axel
    Zottin, Silvia
    Piciarelli, Claudio
    Foresti, Gian Luca
    Colombi, Emanuela
    [J]. INTERNATIONAL JOURNAL ON DOCUMENT ANALYSIS AND RECOGNITION, 2024,
  • [3] Glioma subtype classification from histopathological images using in-domain and out-of-domain transfer learning: An experimental study
    Despotovic, Vladimir
    Kim, Sang-Yoon
    Hau, Ann-Christin
    Kakoichankava, Aliaksandra
    Klamminger, Gilbert Georg
    Borgmann, Felix Bruno Kleine
    Frauenknecht, Katrin B. M.
    Mittelbronn, Michel
    Nazarov, Petr, V
    [J]. HELIYON, 2024, 10 (05)
  • [4] In-Domain versus Out-of-Domain training for Text-Dependent JFA
    Kenny, Patrick
    Stafylakis, Themos
    Alam, Jahangir
    Ouellet, Pierre
    Kockmann, Marcel
    [J]. 15TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2014), VOLS 1-4, 2014, : 1332 - 1336
  • [5] GAN-BASED OUT-OF-DOMAIN DETECTION USING BOTH IN-DOMAIN AND OUT-OF-DOMAIN SAMPLES
    Liang, Chaojie
    Huang, Peijie
    Lai, Wenbin
    Ruan, Ziheng
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7663 - 7667
  • [6] Food Image Classification: The Benefit of In-Domain Transfer Learning
    Touijer, Larbi
    Pastore, Vito Paolo
    Odone, Francesca
    [J]. IMAGE ANALYSIS AND PROCESSING, ICIAP 2023, PT II, 2023, 14234 : 259 - 269
  • [7] Towards Textual Out-of-Domain Detection Without In-Domain Labels
    Jin, Di
    Gao, Shuyang
    Kim, Seokhwan
    Liu, Yang
    Hakkani-Tur, Dilek
    [J]. IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2022, 30 : 1386 - 1395
  • [8] Combining in-domain and out-of-domain speech data for automatic recognition of disordered speech
    Christensen, H.
    Aniol, M. B.
    Bell, P.
    Green, P.
    Hain, T.
    King, S.
    Swietojanski, P.
    [J]. 14TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2013), VOLS 1-5, 2013, : 3609 - 3612
  • [9] KNN-Contrastive Learning for Out-of-Domain Intent Classification
    Zhou, Yunhua
    Liu, Peiju
    Qiu, Xipeng
    [J]. PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 5129 - 5141
  • [10] Exploiting Out-of-Domain Datasets and Visual Representations for Image Sentiment Classification
    Pournaras, Alexandros
    Gkalelis, Nikolaos
    Galanopoulos, Damianos
    Mezaris, Vasileios
    [J]. 2021 16TH INTERNATIONAL WORKSHOP ON SEMANTIC AND SOCIAL MEDIA ADAPTATION & PERSONALIZATION (SMAP 2021), 2021, : 42 - 47