CSDSE: An efficient design space exploration framework for deep neural network accelerator based on cooperative search

被引:0
|
作者
Feng, Kaijie [1 ]
Fan, Xiaoya [1 ]
An, Jianfeng [1 ]
Wang, Haoyang [1 ]
Li, Chuxi [1 ]
机构
[1] Northwestern Polytech Univ, Sch Comp Sci, 1 Dongxiang Rd, Xian 710000, Shaanxi, Peoples R China
关键词
Design space exploration; Deep neural network accelerator; Design Automation; Reinforcement Learning;
D O I
10.1016/j.neucom.2025.129366
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The design and optimization of deep neural network accelerators necessitates thoughtful consideration of numerous design parameters and various resource/physical constraints that render their design spaces massive in scale and complex in distribution. When faced with these massive and complex design spaces, previous works on design space exploration confront the exploration-exploitation dilemma, struggling to concurrently ensure optimization efficiency and stability. To address the exploration-exploitation dilemma, we present a novel design space exploration method entitled CSDSE. CSDSE implements heterogeneous agents separately accountable for exploration or exploitation to cooperatively search the design space. In order to enable CSDSE to adapt design spaces with various space distributions and expanding scales, we extend CSDSE with mechanism of adaptive agent organization and multi-scale search. Furthermore, we introduce a weighted compact buffer that encourages agents to search in diverse directions and bolsters their global exploration ability. CSDSE is implemented to optimize accelerator design. Compared to former DSE methods, it achieves latency speedups of up to 15.68x and energy-delay-product reductions of up to 16.22x under different constraint scenarios.
引用
收藏
页数:20
相关论文
共 50 条
  • [21] The Design and Implementation of Scalable Deep Neural Network Accelerator Cores
    Sakamoto, Ryuichi
    Takata, Ryo
    Ishii, Jun
    Kondo, Masaaki
    Nakamura, Hiroshi
    Ohkubo, Tetsui
    Kojima, Takuya
    Amano, Hideharu
    2017 IEEE 11TH INTERNATIONAL SYMPOSIUM ON EMBEDDED MULTICORE/MANY-CORE SYSTEMS-ON-CHIP (MCSOC 2017), 2017, : 13 - 20
  • [22] CTT-based Non-Volatile Deep Neural Network Accelerator Design
    Xiao, Yang
    Fan, Wuyu
    Du, Yuan
    Du, Li
    Chang, Mau-Chung Frank
    18TH INTERNATIONAL SOC DESIGN CONFERENCE 2021 (ISOCC 2021), 2021, : 258 - 259
  • [23] Guiding Exploration of Antimicrobial Peptide Space with a Deep Neural Network
    Dua, Manpriya
    Veltri, Daniel
    Bishop, Barney
    Shehu, Amarda
    PROCEEDINGS 2018 IEEE INTERNATIONAL CONFERENCE ON BIOINFORMATICS AND BIOMEDICINE (BIBM), 2018, : 2082 - 2087
  • [24] Activation in Network for NoC-based Deep Neural Network Accelerator
    Zhu, Wenyao
    Chen, Yizhi
    Lu, Zhonghai
    2024 INTERNATIONAL VLSI SYMPOSIUM ON TECHNOLOGY, SYSTEMS AND APPLICATIONS, VLSI TSA, 2024,
  • [25] Software-Defined Design Space Exploration for an Efficient DNN Accelerator Architecture
    Yu, Ye
    Li, Yingmin
    Che, Shuai
    Jha, Niraj K.
    Zhang, Weifeng
    IEEE TRANSACTIONS ON COMPUTERS, 2021, 70 (01) : 45 - 56
  • [26] Performance Modeling of Stencil Computing on a Stream-Based FPGA Accelerator for Efficient Design Space Exploration
    Dohi, Keisuke
    Okina, Koji
    Soejima, Rie
    Shibata, Yuichiro
    Oguri, Kiyoshi
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2015, E98D (02): : 298 - 308
  • [27] Exploring the Design Space of Efficient Deep Neural Networks
    Yu, Fuxun
    Stamoulis, Dimitrios
    Wang, Di
    Lymberopoulos, Dimitrios
    Chen, Xiang
    2020 IEEE/ACM SYMPOSIUM ON EDGE COMPUTING (SEC 2020), 2020, : 317 - 318
  • [28] Cooperative Initialization based Deep Neural Network Training
    Singh, Pravendra
    Varshney, Munender
    Namboodiri, Vinay P.
    2020 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2020, : 1130 - 1139
  • [29] An Optical Accelerator for Deep Neural Network Based on Integrated Nanophotonics
    Shiomi, Jun
    Ishihara, Tohru
    Onodera, Hidetoshi
    Shinya, Akihiko
    Notomi, Masaya
    2020 INTERNATIONAL CONFERENCE ON REBOOTING COMPUTING (ICRC 2020), 2020, : 95 - 101
  • [30] A Deep Neural Network Accelerator Based on Tiled RRAM Architecture
    Wang, Qiwen
    Wang, Xinxin
    Lee, Seung Hwan
    Meng, Fan-Hsuan
    Lu, Wei D.
    2019 IEEE INTERNATIONAL ELECTRON DEVICES MEETING (IEDM), 2019,