Title :
SNNAP: Approximate computing on programmable SoCs via neural acceleration
Author :
Moreau, Thierry ; Wyse, Mark ; Nelson, Jacob ; Sampson, Adrian ; Esmaeilzadeh, Hadi ; Ceze, Luis ; Oskin, Mark
Author_Institution :
Univ. of Washington, Seattle, WA, USA
Abstract :
Many applications that can take advantage of accelerators are amenable to approximate execution. Past work has shown that neural acceleration is a viable way to accelerate approximate code. In light of the growing availability of on-chip field-programmable gate arrays (FPGAs), this paper explores neural acceleration on off-the-shelf programmable SoCs. We describe the design and implementation of SNNAP, a flexible FPGA-based neural accelerator for approximate programs. SNNAP is designed to work with a compiler workflow that configures the neural network´s topology and weights instead of the programmable logic of the FPGA itself. This approach enables effective use of neural acceleration in commercially available devices and accelerates different applications without costly FPGA reconfigurations. No hardware expertise is required to accelerate software with SNNAP, so the effort required can be substantially lower than custom hardware design for an FPGA fabric and possibly even lower than current “C-to-gates” high-level synthesis (HLS) tools. Our measurements on a Xilinx Zynq FPGA show that SNNAP yields a geometric mean of 3.8× speedup (as high as 38.1×) and 2.8× energy savings (as high as 28 x) with less than 10% quality loss across all applications but one. We also compare SNNAP with designs generated by commercial HLS tools and show that SNNAP has similar performance overall, with better resource-normalized throughput on 4 out of 7 benchmarks.
Keywords :
electronic engineering computing; field programmable gate arrays; high level synthesis; logic design; network topology; neural nets; programmable logic devices; system-on-chip; C-to-gates high-level synthesis tools; FPGA fabric; FPGA reconfigurations; FPGA-based neural accelerator; HLS tools; SNNAP; Xilinx Zynq FPGA; approximate code; approximate computing; approximate programs; compiler workflow; energy savings; hardware design; neural acceleration; neural network topology; on-chip field-programmable gate arrays; programmable SoC; programmable logic; quality loss; systolic neural network accelerator; Acceleration; Biological neural networks; Field programmable gate arrays; Neurons; System-on-chip; Topology;
Conference_Titel :
High Performance Computer Architecture (HPCA), 2015 IEEE 21st International Symposium on
Conference_Location :
Burlingame, CA
DOI :
10.1109/HPCA.2015.7056066