SparseACC: A Generalized Linear Model Accelerator for Sparse Datasets

被引:0
|
作者
Zhang, Jie [1 ]
Huang, Hongjing [1 ]
Sun, Jie [1 ]
Luna, Juan Gomez [2 ]
Mutlu, Onur [2 ]
Wang, Zeke [3 ]
机构
[1] Zhejiang Univ, Dept Comp Sci, Hangzhou 310058, Peoples R China
[2] Swiss Fed Inst Technol, Dept Comp Sci, CH-8092 Zurich, Switzerland
[3] Zhejiang Univ, Dept Comp Sci, Hangzhou 310027, Peoples R China
关键词
Accelerator; linear model; sparse; training;
D O I
10.1109/TCAD.2023.3324276
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Stochastic gradient descent (SGD) is widely used for training generalized linear models (GLMs), such as support vector machine and logistic regression, on large industry datasets. Such a training consumes plenty of computing power and therefore plenty of accelerators are proposed to accelerate the GLM training. However, real-world datasets are always highly sparse. For example, YouTube's social network connectivity contains only 2.31% nonzero elements (NZs). It is not trivial to design an accelerator that is able to efficiently train on a sparse dataset that is stored in a compressed sparse format (e.g., compressed sparse row (CSR) format). The design of such an accelerator faces three challenges: 1) bank conflicts, which may happen when multiple processing engines in the accelerator access multiple memory banks; 2) complex interconnections, which are necessary to allow all processing engines to access any memory bank; and 3) high-synchronization overhead, since each sample in sparse dataset has a different number of NZs and these elements have different distributions, thus it is hard to overlap gradient computation and model update of neighboring batches. To this end, we propose SparseACC, a sparsity-aware accelerator for training generalized linear models (GLMs). SparseACC is based on two key mechanisms. First, a software/hardware co-design approach solves the first two design challenges by proposing a novel bank-conflict-free (BCF) and bank-balanced CSR format. Second, a weight-aware ping-pong model solves the third challenge, thus maximizing the utilization of the processing engines. SparseACC leverages these two mechanisms to orchestrate training over sparse datasets, such that the training time decreases linearly with the sparsity of the dataset. We prototype SparseACC on a Xilinx Alveo U280 FPGA (Xilinx, 2020). The experimental evaluation shows that SparseACC converges up to 3.5 x , 18 x , 38 x , and 110x faster than the state-of-the-art counterparts on a sparse accelerator, a Tesla V100 GPU, an Intel i9-10900k CPU, and a dense accelerator, respectively.
引用
收藏
页码:840 / 853
页数:14
相关论文
共 50 条
  • [1] A sparse linear regression model for incomplete datasets
    Marcelo B. A. Veras
    Diego P. P. Mesquita
    Cesar L. C. Mattos
    João P. P. Gomes
    Pattern Analysis and Applications, 2020, 23 : 1293 - 1303
  • [2] A sparse linear regression model for incomplete datasets
    Veras, Marcelo B. A.
    Mesquita, Diego P. P.
    Mattos, Cesar L. C.
    Gomes, Joao P. P.
    PATTERN ANALYSIS AND APPLICATIONS, 2020, 23 (03) : 1293 - 1303
  • [3] Multiplicative Weight for Sparse Generalized Linear Model
    Cai, Qianlong
    Wang, Ziyiyang
    Xie, Shuting
    Deng, Siting
    2021 3RD INTERNATIONAL CONFERENCE ON MACHINE LEARNING, BIG DATA AND BUSINESS INTELLIGENCE (MLBDBI 2021), 2021, : 245 - 248
  • [4] Robust and sparse regression in generalized linear model by stochastic optimization
    Takayuki Kawashima
    Hironori Fujisawa
    Japanese Journal of Statistics and Data Science, 2019, 2 : 465 - 489
  • [5] Robust and sparse regression in generalized linear model by stochastic optimization
    Kawashima, Takayuki
    Fujisawa, Hironori
    JAPANESE JOURNAL OF STATISTICS AND DATA SCIENCE, 2019, 2 (02) : 465 - 489
  • [6] Using the dglars Package to Estimate a Sparse Generalized Linear Model
    Augugliaro, Luigi
    Mineo, Angelo M.
    Advances in Statistical Models for Data Analysis, 2015, : 1 - 8
  • [7] Estimating overdispersion when fitting a generalized linear model to sparse data
    Fletcher, D. J.
    BIOMETRIKA, 2012, 99 (01) : 230 - 237
  • [8] ZeD: A Generalized Accelerator for Variably Sparse Matrix Computations in ML
    Dangi, Pranav
    Bai, Zhenyu
    Juneja, Rohan
    Wijerathne, Dhananjaya
    Mitra, Tulika
    PROCEEDINGS OF THE 2024 THE INTERNATIONAL CONFERENCE ON PARALLEL ARCHITECTURES AND COMPILATION TECHNIQUES, PACT 2024, 2024, : 246 - 257
  • [9] SPARSE GENERALIZED FUNCTIONAL LINEAR MODEL FOR PREDICTING REMISSION STATUS OF DEPRESSION PATIENTS
    Liu, Yashu
    Nie, Zhi
    Zhou, Jiayu
    Farnum, Michael
    Narayan, Vaibhav A.
    Wittenberg, Gayle
    Ye, Jieping
    PACIFIC SYMPOSIUM ON BIOCOMPUTING 2014, 2014, : 364 - 375
  • [10] MODEL OF A LINEAR-ACCELERATOR
    MASS, ND
    AMERICAN JOURNAL OF PHYSICS, 1975, 43 (03) : 277 - 278