DocumentCode :
587604
Title :
BenchNN: On the broad potential application scope of hardware neural network accelerators
Author :
Tianshi Chen ; Yunji Chen ; Duranton, Marc ; Qi Guo ; Hashmi, A. ; Lipasti, M. ; Nere, A. ; Shi Qiu ; Sebag, Michele ; Temam, Olivier
Author_Institution :
ICT, China
fYear :
2012
fDate :
4-6 Nov. 2012
Firstpage :
36
Lastpage :
45
Abstract :
Recent technology trends have indicated that, although device sizes will continue to scale as they have in the past, supply voltage scaling has ended. As a result, future chips can no longer rely on simply increasing the operational core count to improve performance without surpassing a reasonable power budget. Alternatively, allocating die area towards accelerators targeting an application, or an application domain, appears quite promising, and this paper makes an argument for a neural network hardware accelerator. After being hyped in the 1990s, then fading away for almost two decades, there is a surge of interest in hardware neural networks because of their energy and fault-tolerance properties. At the same time, the emergence of high-performance applications like Recognition, Mining, and Synthesis (RMS) suggest that the potential application scope of a hardware neural network accelerator would be broad. In this paper, we want to highlight that a hardware neural network accelerator is indeed compatible with many of the emerging high-performance workloads, currently accepted as benchmarks for high-performance micro-architectures. For that purpose, we develop and evaluate software neural network implementations of 5 (out of 12) RMS applications from the PARSEC Benchmark Suite. Our results show that neural network implementations can achieve competitive results, with respect to application-specific quality metrics, on these 5 RMS applications.
Keywords :
neural chips; neural net architecture; power aware computing; software fault tolerance; BenchNN; PARSEC Benchmark Suite; RMS applications; application domain; application-specific quality metrics; device sizes; die area allocation; energy properties; fault-tolerance properties; hardware neural network accelerators; high-performance microarchitectures; high-performance workloads; recognition-mining-and-synthesis; software neural network implementations; voltage scaling; Artificial neural networks; Benchmark testing; Hardware; Neurons; Optimization; Wires; PARSEC; accelerator; benchmark; neural networks;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Workload Characterization (IISWC), 2012 IEEE International Symposium on
Conference_Location :
La Jolla, CA
Print_ISBN :
978-1-4673-4531-6
Type :
conf
DOI :
10.1109/IISWC.2012.6402898
Filename :
6402898
Link To Document :
بازگشت