A transformer-based neural ODE for dense prediction

被引:0
|
作者
Seyedalireza Khoshsirat
Chandra Kambhamettu
机构
[1] University of Delaware,Department of Computer and Information Sciences
来源
关键词
Neural ODE; Dense prediction; Transformer;
D O I
暂无
中图分类号
学科分类号
摘要
Neural ordinary differential equations (ODEs) represent an emergent class of deep learning models exhibiting continuous depth. While they have shown promising results across various machine learning tasks, existing methods for dense prediction tasks have not fully harnessed their potential, often due to employing sub-optimal architectural components and limited dataset studies. To address this, our paper introduces a robust neural ODE architecture specifically tailored for dense prediction tasks and performs an extensive evaluation across a broad range of datasets. Our approach draws upon proven design elements from top-performing networks, integrating transformer blocks as core building blocks. Unique to our design is the retention of multiple concurrent representations at varying resolutions throughout the network. These representations continually exchange information, ensuring they remain updated. Our network achieves unrivaled performance in tasks such as image classification, semantic segmentation, and answer grounding. We conduct several ablation studies to shed light on the impacts of various design parameters. Our results affirm the effectiveness of our approach and its potential for further advancements in dense prediction tasks.
引用
收藏
相关论文
共 50 条
  • [1] A transformer-based neural ODE for dense prediction
    Khoshsirat, Seyedalireza
    Kambhamettu, Chandra
    [J]. MACHINE VISION AND APPLICATIONS, 2023, 34 (06)
  • [2] A Neural ODE and Transformer-based Model for Temporal Understanding and Dense Video Captioning
    Artham, Sainithin
    Shaikh, Soharab Hossain
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (23) : 64037 - 64056
  • [3] RPConvformer: A novel Transformer-based deep neural networks for traffic flow prediction
    Wen, Yanjie
    Xu, Ping
    Li, Zhihong
    Xu, Wangtu
    Wang, Xiaoyu
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2023, 218
  • [4] A transformer-based neural network framework for full names prediction with abbreviations and contexts
    Ye, Ziming
    Li, Shuangyin
    [J]. DATA & KNOWLEDGE ENGINEERING, 2024, 150
  • [5] A Transformer-based Neural Architecture Search Method
    Wang, Shang
    Tang, Huanrong
    Ouyang, Jianquan
    [J]. PROCEEDINGS OF THE 2023 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE COMPANION, GECCO 2023 COMPANION, 2023, : 691 - 694
  • [6] Transformer-based structural seismic response prediction
    Zhang, Qingyu
    Guo, Maozi
    Zhao, Lingling
    Li, Yang
    Zhang, Xinxin
    Han, Miao
    [J]. STRUCTURES, 2024, 61
  • [7] Temporal fusion transformer-based prediction in aquaponics
    Metin, Ahmet
    Kasif, Ahmet
    Catal, Cagatay
    [J]. JOURNAL OF SUPERCOMPUTING, 2023, 79 (17): : 19934 - 19958
  • [8] A Transformer-Based Framework for Geomagnetic Activity Prediction
    Abduallah, Yasser
    Wang, Jason T. L.
    Xu, Chunhui
    Wang, Haimin
    [J]. FOUNDATIONS OF INTELLIGENT SYSTEMS (ISMIS 2022), 2022, 13515 : 325 - 335
  • [9] Transformer-based Neural Network for Electrocardiogram Classification
    Atiea, Mohammed A.
    Adel, Mark
    [J]. INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2022, 13 (11) : 357 - 363
  • [10] Privacy Protection in Transformer-based Neural Network
    Lang, Jiaqi
    Li, Linjing
    Chen, Weiyun
    Zeng, Daniel
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON INTELLIGENCE AND SECURITY INFORMATICS (ISI), 2019, : 182 - 184