Software/Hardware Co-design for Multi-modal Multi-task Learning in Autonomous Systems

被引:13
|
作者
Hao, Cong [1 ]
Chen, Deming [2 ]
机构
[1] Georgia Inst Technol, Atlanta, GA 30332 USA
[2] Univ Illinois, Urbana, IL USA
关键词
D O I
10.1109/AICAS51828.2021.9458577
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Optimizing the quality of result (QoR) and the quality of service (QoS) of AI-empowered autonomous systems simultaneously is very challenging. First, there are multiple input sources, e.g., multi-modal data from different sensors, requiring diverse data preprocessing, sensor fusion, and feature aggregation. Second, there are multiple tasks that require various AI models to run simultaneously, e.g., perception, localization, and control. Third, the computing and control system is heterogeneous, composed of hardware components with varied features, such as embedded CPUs, GPUs, FPGAs, and dedicated accelerators. Therefore, autonomous systems essentially require multi-modal multi-task (MMMT) learning which must be aware of hardware performance and implementation strategies. While MMMT learning has been attracting intensive research interests, its applications in autonomous systems are still underexplored. In this paper, we first discuss the opportunities of applying MMMT techniques in autonomous systems, and then discuss the unique challenges that must be solved. In addition, we discuss the necessity and opportunities of MMMT model and hardware co-design, which is critical for autonomous systems especially with power/resource-limited or heterogeneous platforms. We formulate the MMMT model and heterogeneous hardware implementation co-design as a differentiable optimization problem, with the objective of improving the solution quality and reducing the overall power consumption and critical path latency. We advocate for further explorations of MMMT in autonomous systems and software/hardware co-design solutions.
引用
收藏
页数:5
相关论文
共 50 条
  • [31] Gaining Extra Supervision via Multi-task learning for Multi-Modal Video Question Answering
    Kim, Junyeong
    Ma, Minuk
    Kim, Kyungsu
    Kim, Sungjin
    Yoo, Chang D.
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [32] Multi-modal Sentiment and Emotion Joint Analysis with a Deep Attentive Multi-task Learning Model
    Zhang, Yazhou
    Rong, Lu
    Li, Xiang
    Chen, Rui
    ADVANCES IN INFORMATION RETRIEVAL, PT I, 2022, 13185 : 518 - 532
  • [33] Multi-task Classification Model Based On Multi-modal Glioma Data
    Li, Jialun
    Jin, Yuanyuan
    Yu, Hao
    Wang, Xiaoling
    Zhuang, Qiyuan
    Chen, Liang
    11TH IEEE INTERNATIONAL CONFERENCE ON KNOWLEDGE GRAPH (ICKG 2020), 2020, : 165 - 172
  • [34] Multi-task Learning of Semantic Segmentation and Height Estimation for Multi-modal Remote Sensing Images
    Mengyu WANG
    Zhiyuan YAN
    Yingchao FENG
    Wenhui DIAO
    Xian SUN
    Journal of Geodesy and Geoinformation Science, 2023, 6 (04) : 27 - 39
  • [35] STARS: Soft Multi-Task Learning for Activity Recognition from Multi-Modal Sensor Data
    Liu, Xi
    Tan, Pang-Ning
    Liu, Lei
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2018, PT II, 2018, 10938 : 569 - 581
  • [36] MULTI-MODAL MULTI-TASK LEARNING FOR SEMANTIC SEGMENTATION OF LAND COVER UNDER CLOUDY CONDITIONS
    Xu, Fang
    Shi, Yilei
    Yang, Wen
    Zhu, Xiaoxiang
    IGARSS 2023 - 2023 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM, 2023, : 6274 - 6277
  • [37] A multi-modal fusion framework based on multi-task correlation learning for cancer prognosis prediction
    Tan, Kaiwen
    Huang, Weixian
    Liu, Xiaofeng
    Hu, Jinlong
    Dong, Shoubin
    ARTIFICIAL INTELLIGENCE IN MEDICINE, 2022, 126
  • [38] Multi-Modal Multi-Task Learning for Joint Prediction of Clinical Scores in Alzheimer's Disease
    Zhang, Daoqiang
    Shen, Dinggang
    MULTIMODAL BRAIN IMAGE ANALYSIS, 2011, 7012 : 60 - 67
  • [39] Fast Multi-Task SCCA Learning with Feature Selection for Multi-Modal Brain Imaging Genetics
    Du, Lei
    Liu, Kefei
    Yao, Xiaohui
    Risacher, Shannon L.
    Han, Junwei
    Guo, Lei
    Saykin, Andrew J.
    Shen, Li
    Weiner, Michael
    Aisen, Paul
    Petersen, Ronald
    Jack, Clifford R., Jr.
    Jagust, William
    Trojanowki, John Q.
    Toga, Arthur W.
    Beckett, Laurel
    Green, Robert C.
    Saykin, Andrew J.
    Morris, John
    Liu, Enchi
    Montine, Tom
    Gamst, Anthony
    Thomas, Ronald G.
    Donohue, Michael
    Walter, Sarah
    Gessert, Devon
    Sather, Tamie
    Harvey, Danielle
    Kornak, John
    Dale, Anders
    Bernstein, Matthew
    Felmlee, Joel
    Fox, Nick
    Thompson, Paul
    Schuff, Norbert
    Alexander, Gene
    DeCarli, Charles
    Bandy, Dan
    Koeppe, Robert A.
    Foster, Norm
    Reiman, Eric M.
    Chen, Kewei
    Mathis, Chet
    Cairns, Nigel J.
    Taylor-Reinwald, Lisa
    Shaw, Les
    Lee, Virginia M. Y.
    Korecka, Magdalena
    Crawford, Karen
    Neu, Scott
    PROCEEDINGS 2018 IEEE INTERNATIONAL CONFERENCE ON BIOINFORMATICS AND BIOMEDICINE (BIBM), 2018, : 356 - 361
  • [40] VEMO: A Versatile Elastic Multi-modal Model for Search-Oriented Multi-task Learning
    Fei, Nanyi
    Jiang, Hao
    Lu, Haoyu
    Long, Jinqiang
    Dai, Yanqi
    Fan, Tuo
    Cao, Zhao
    Lu, Zhiwu
    ADVANCES IN INFORMATION RETRIEVAL, ECIR 2024, PT I, 2024, 14608 : 56 - 72