• DocumentCode
    62971
  • Title

    A High-Throughput Neural Network Accelerator

  • Author

    Tianshi Chen ; Zidong Du ; Ninghui Sun ; Jia Wang ; Chengyong Wu ; Yunji Chen ; Temam, Olivier

  • Volume
    35
  • Issue
    3
  • fYear
    2015
  • fDate
    May-June 2015
  • Firstpage
    24
  • Lastpage
    32
  • Abstract
    The authors designed an accelerator architecture for large-scale neural networks, with an emphasis on the impact of memory on accelerator design, performance, and energy. In this article, they present a concrete design at 65 nm that can perform 496 16-bit fixed-point operations in parallel every 1.02 ns, that is, 452 gop/s, in a 3.02mm2, 485-mw footprint (excluding main memory accesses).
  • Keywords
    neural nets; accelerator architecture; accelerator design; high-throughput neural network accelerator; size 65 nm; Accelerators; Artificial neural networks; Computer architecture; Graphics processing units; Machine learning; Market research; Neural networks; hardware accelerator; machine learning; neural network;
  • fLanguage
    English
  • Journal_Title
    Micro, IEEE
  • Publisher
    ieee
  • ISSN
    0272-1732
  • Type

    jour

  • DOI
    10.1109/MM.2015.41
  • Filename
    7106400