A Transprecision Floating-Point Platform for Ultra-Low Power Computing

被引:0
|
作者
Tagliavini, Giuseppe [1 ]
Mach, Stefan [2 ]
Rossi, Davide [1 ]
Marongiu, Andrea [3 ]
Benini, Luca [1 ,2 ]
机构
[1] Univ Bologna, DEI, Bologna, Italy
[2] Swiss Fed Inst Technol, IIS, Zurich, Switzerland
[3] Univ Bologna, DISI, Bologna, Italy
基金
欧盟地平线“2020”;
关键词
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In modern low-power embedded platforms, the execution of floating-point (FP) operations emerges as a major contributor to the energy consumption of compute-intensive applications with large dynamic range. Experimental evidence shows that 50% of the energy consumed by a core and its data memory is related to FP computations. The adoption of FP formats requiring a lower number of bits is an interesting opportunity to reduce energy consumption, since it allows to simplify the arithmetic circuitry and to reduce the memory bandwidth required to transfer data between memory and registers by enabling vectorization. From a theoretical point of view, the adoption of multiple FP types perfectly fits with the principle of transprecision computing, allowing fine-grained control of approximation while meeting specified constraints on the precision of final results. In this paper we propose an extended FP type system with complete hardware support to enable transprecision computing on low-power embedded processors, including two standard formats (binary32 and binary16) and two new formats (binary8 and binary16alt). First, we introduce a software library that enables exploration of FP types by tuning both precision and dynamic range of program variables. Then, we present a methodology to integrate our library with an external tool for precision tuning, and experimental results that highlight the clear benefits of introducing the new formats. Finally, we present the design of a transprecision FP unit capable of handling 8-bit and 16-hit operations in addition to standard 32 bit operations. Experimental results on FP-intensive benchmarks show that up to 90% of FP operations can be safely scaled down to 8-bit or 16-bit formats. Thanks to precision tuning and vectorization, execution time is decreased by 12% and memory accesses are reduced by 27% on average, leading to a reduction of energy consumption up to 30%.
引用
收藏
页码:1051 / 1056
页数:6
相关论文
共 50 条
  • [1] A Transprecision Floating-Point Architecture for Energy-Efficient Embedded Computing
    Mach, Stefan
    Rossi, Davide
    Tagliavini, Giuseppe
    Marongiu, Andrea
    Benini, Luca
    2018 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2018,
  • [2] Towards a Transprecision Polymorphic Floating-Point Unit for Mixed-precision Computing
    Linhares, Alisson
    Azevedo, Rodolfo
    2019 31ST INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE AND HIGH PERFORMANCE COMPUTING (SBAC-PAD 2019), 2019, : 56 - 63
  • [3] A Low-Power Transprecision Floating-Point Cluster for Efficient Near-Sensor Data Analytics
    Montagna, Fabio
    Mach, Stefan
    Benatti, Simone
    Garofalo, Angelo
    Ottavi, Gianmarco
    Benini, Luca
    Rossi, Davide
    Tagliavini, Giuseppe
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2022, 33 (05) : 1038 - 1053
  • [4] Floating Point CGRA based Ultra-Low Power DSP Accelerator
    Rohit Prasad
    Satyajit Das
    Kevin J. M. Martin
    Philippe Coussy
    Journal of Signal Processing Systems, 2021, 93 : 1159 - 1171
  • [5] Floating Point CGRA based Ultra-Low Power DSP Accelerator
    Prasad, Rohit
    Das, Satyajit
    Martin, Kevin J. M.
    Coussy, Philippe
    JOURNAL OF SIGNAL PROCESSING SYSTEMS FOR SIGNAL IMAGE AND VIDEO TECHNOLOGY, 2021, 93 (10): : 1159 - 1171
  • [6] TRANSPIRE: An energy-efficient TRANSprecision floating-point Programmable archItectuRE
    Prasad, Rohit
    Das, Satyajit
    Martin, Kevin J. M.
    Tagliavini, Giuseppe
    Coussy, Philippe
    Benini, Luca
    Rossi, Davide
    PROCEEDINGS OF THE 2020 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE 2020), 2020, : 1067 - 1072
  • [7] FPnew: An Open-Source Multiformat Floating-Point Unit Architecture for Energy-Proportional Transprecision Computing
    Mach, Stefan
    Schuiki, Fabian
    Zaruba, Florian
    Benini, Luca
    IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2021, 29 (04) : 774 - 787
  • [8] Design of low power approximate floating-point adders
    Omidi, Reza
    Sharifzadeh, Sepehr
    INTERNATIONAL JOURNAL OF CIRCUIT THEORY AND APPLICATIONS, 2021, 49 (01) : 185 - 195
  • [9] Energy-Efficient Vision on the PULP Platform for Ultra-Low Power Parallel Computing
    Conti, Francesco
    Rossi, Davide
    Pullini, Antonio
    Loi, Igor
    Benini, Luca
    PROCEEDINGS OF THE 2014 IEEE WORKSHOP ON SIGNAL PROCESSING SYSTEMS (SIPS 2014), 2014, : 274 - 279
  • [10] Accurate and Reliable Computing in Floating-Point Arithmetic
    Rump, Siegfried M.
    MATHEMATICAL SOFTWARE - ICMS 2010, 2010, 6327 : 105 - 108