DocumentCode :
245393
Title :
Training itself: Mixed-signal training acceleration for memristor-based neural network
Author :
Boxun Li ; Yuzhi Wang ; Yu Wang ; Chen, Yuanfeng ; Huazhong Yang
Author_Institution :
Dept. of E.E., Tsinghua Univ., Beijing, China
fYear :
2014
fDate :
20-23 Jan. 2014
Firstpage :
361
Lastpage :
366
Abstract :
The artificial neural network (ANN) is among the most widely used methods in data processing applications. The memristor-based neural network further demonstrates a power efficient hardware realization of ANN. Training phase is the critical operation of memristor-based neural network. However, the traditional training method for memristor-based neural network is time consuming and energy inefficient. Users have to first work out the parameters of memristors through digital computing systems and then tune the memristor to the corresponding state. In this work, we introduce a mixed-signal training acceleration framework, which realizes the self-training of memristor-based neural network. We first modify the original stochastic gradient descent algorithm by approximating calculations and designing an alternative computing method. We then propose a mixed-signal acceleration architecture for the modified training algorithm by equipping the original memristor-based neural network architecture with the copy crossbar technique, weight update units, sign calculation units and other assistant units. The experiment on the MNIST database demonstrates that the proposed mixed-signal acceleration is 3 orders of magnitude faster and 4 orders of magnitude more energy efficient than the CPU implementation counterpart at the cost of a slight decrease of the recognition accuracy (<; 5%).
Keywords :
gradient methods; learning (artificial intelligence); memristors; neural nets; stochastic processes; artificial neural network; digital computing systems; memristor-based neural network; mixed-signal training acceleration; power efficient hardware realization; stochastic gradient descent algorithm; training phase; Acceleration; Arrays; Convergence; Memristors; Mirrors; Neural networks; Training; Memristor; Neural Network; Training;
fLanguage :
English
Publisher :
ieee
Conference_Titel :
Design Automation Conference (ASP-DAC), 2014 19th Asia and South Pacific
Conference_Location :
Singapore
Type :
conf
DOI :
10.1109/ASPDAC.2014.6742916
Filename :
6742916
Link To Document :
بازگشت