MDL-NAS: A Joint Multi-domain Learning Framework for Vision Transformer

被引:8
|
作者
Wang, Shiguang [1 ,3 ]
Xie, Tao [2 ,3 ]
Cheng, Jian [1 ]
Zhang, Xingcheng [3 ]
Liu, Haijun [4 ]
机构
[1] Univ Elect Sci & Technol China, Chengdu, Peoples R China
[2] Harbin Inst Technol, Harbin, Peoples R China
[3] SenseTime Res, Hong Kong, Peoples R China
[4] Chongqing Univ, Chongqing, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/CVPR52729.2023.01924
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this work, we introduce MDL-NAS, a unified framework that integrates multiple vision tasks into a manageable supernet and optimizes these tasks collectively under diverse dataset domains. MDL-NAS is storage-efficient since multiple models with a majority of shared parameters can be deposited into a single one. Technically, MDL-NAS constructs a coarse-to-fine search space, where the coarse search space offers various optimal architectures for different tasks while the fine search space provides fine-grained parameter sharing to tackle the inherent obstacles of multi-domain learning. In the fine search space, we suggest two parameter sharing policies, i.e., sequential sharing policy and mask sharing policy. Compared with previous works, such two sharing policies allow for the partial sharing and non-sharing of parameters at each layer of the network, hence attaining real fine-grained parameter sharing. Finally, we present a joint-subnet search algorithm that finds the optimal architecture and sharing parameters for each task within total resource constraints, challenging the traditional practice that downstream vision tasks are typically equipped with backbone networks designed for image classification. Experimentally, we demonstrate that MDL-NAS families fitted with non-hierarchical or hierarchical transformers deliver competitive performance for all tasks compared with state-of-the-art methods while maintaining efficient storage deployment and computation. We also demonstrate that MDL-NAS allows incremental learning and evades catastrophic forgetting when generalizing to a new task.
引用
收藏
页码:20094 / 20104
页数:11
相关论文
共 50 条
  • [32] A CONTROL FRAMEWORK OF SURVIVABLE MULTI-DOMAIN OPTICAL NETWORKS
    He, He
    Mouftah, Hussein
    Wu, Jing
    2011 24TH CANADIAN CONFERENCE ON ELECTRICAL AND COMPUTER ENGINEERING (CCECE), 2011, : 54 - 57
  • [33] Integrating AADL within a multi-domain modeling framework
    Malavolta, Ivano
    Muccini, Henry
    Pelliccione, Patrizio
    2009 14TH IEEE INTERNATIONAL CONFERENCE ON ENGINEERING OF COMPLEX COMPUTER SYSTEMS (ICECCS), 2009, : 342 - 347
  • [34] Unsupervised multi-domain image translation with domain representation learning
    Liu, Huajun
    Chen, Lei
    Sui, Haigang
    Zhu, Qing
    Lei, Dian
    Liu, Shubo
    SIGNAL PROCESSING-IMAGE COMMUNICATION, 2021, 99
  • [35] Multi-Domain Sequential Recommendation via Domain Space Learning
    Hwang, Junyoung
    Ju, Hyunjun
    Kang, SeongKu
    Jang, Sanghwan
    Yu, Hwanjo
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 2134 - 2144
  • [36] AI and ML in the Multi-Domain Operations Era: Vision and Pitfalls
    Baker, Michael A.
    Al-Khalifa, Khaled A.
    Harlas, Ioannis N.
    King, Marvin L.
    ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING FOR MULTI-DOMAIN OPERATIONS APPLICATIONS II, 2020, 11413
  • [37] Joint learning of images and videos with a single Vision Transformer
    Shimizu, Shuki
    Tamaki, Toru
    2023 18TH INTERNATIONAL CONFERENCE ON MACHINE VISION AND APPLICATIONS, MVA, 2023,
  • [38] Joint Multi-Domain Channel Estimation Based on Sparse Bayesian Learning for OTFS System br
    Liao, Yong
    Li, Xue
    CHINA COMMUNICATIONS, 2023, 20 (01) : 14 - 23
  • [39] Multi-Domain Multi-Task Rehearsal for Lifelong Learning
    Lyu, Fan
    Wang, Shuai
    Feng, Wei
    Ye, Zihan
    Hu, Fuyuan
    Wang, Song
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 8819 - 8827
  • [40] Scalarization for Multi-Task and Multi-Domain Learning at Scale
    Royer, Amelie
    Blankevoort, Tijmen
    Bejnordi, Babak Ehteshami
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,