• DocumentCode
    245393
  • Title

    Training itself: Mixed-signal training acceleration for memristor-based neural network

  • Author

    Boxun Li ; Yuzhi Wang ; Yu Wang ; Chen, Yuanfeng ; Huazhong Yang

  • Author_Institution
    Dept. of E.E., Tsinghua Univ., Beijing, China
  • fYear
    2014
  • fDate
    20-23 Jan. 2014
  • Firstpage
    361
  • Lastpage
    366
  • Abstract
    The artificial neural network (ANN) is among the most widely used methods in data processing applications. The memristor-based neural network further demonstrates a power efficient hardware realization of ANN. Training phase is the critical operation of memristor-based neural network. However, the traditional training method for memristor-based neural network is time consuming and energy inefficient. Users have to first work out the parameters of memristors through digital computing systems and then tune the memristor to the corresponding state. In this work, we introduce a mixed-signal training acceleration framework, which realizes the self-training of memristor-based neural network. We first modify the original stochastic gradient descent algorithm by approximating calculations and designing an alternative computing method. We then propose a mixed-signal acceleration architecture for the modified training algorithm by equipping the original memristor-based neural network architecture with the copy crossbar technique, weight update units, sign calculation units and other assistant units. The experiment on the MNIST database demonstrates that the proposed mixed-signal acceleration is 3 orders of magnitude faster and 4 orders of magnitude more energy efficient than the CPU implementation counterpart at the cost of a slight decrease of the recognition accuracy (<; 5%).
  • Keywords
    gradient methods; learning (artificial intelligence); memristors; neural nets; stochastic processes; artificial neural network; digital computing systems; memristor-based neural network; mixed-signal training acceleration; power efficient hardware realization; stochastic gradient descent algorithm; training phase; Acceleration; Arrays; Convergence; Memristors; Mirrors; Neural networks; Training; Memristor; Neural Network; Training;
  • fLanguage
    English
  • Publisher
    ieee
  • Conference_Titel
    Design Automation Conference (ASP-DAC), 2014 19th Asia and South Pacific
  • Conference_Location
    Singapore
  • Type

    conf

  • DOI
    10.1109/ASPDAC.2014.6742916
  • Filename
    6742916