skip to main content
Primo Search
Search in: Busca Geral

SCNN: An accelerator for compressed-sparse convolutional neural networks

Parashar, Angshuman ; Minsoo Rhu ; Mukkara, Anurag ; Puglielli, Antonio ; Venkatesan, Rangharajan ; Khailany, Brucek ; Emer, Joel ; Keckler, Stephen W. ; Dally, William J.

2017 ACM/IEEE 44th Annual International Symposium on Computer Architecture (ISCA), 2017, p.27-40

ACM

Sem texto completo

Citações Citado por
  • Título:
    SCNN: An accelerator for compressed-sparse convolutional neural networks
  • Autor: Parashar, Angshuman ; Minsoo Rhu ; Mukkara, Anurag ; Puglielli, Antonio ; Venkatesan, Rangharajan ; Khailany, Brucek ; Emer, Joel ; Keckler, Stephen W. ; Dally, William J.
  • Assuntos: accelerator architecture ; Accelerator architectures ; Arrays ; Convolutional neural networks ; Encoding ; Random access memory ; Training
  • É parte de: 2017 ACM/IEEE 44th Annual International Symposium on Computer Architecture (ISCA), 2017, p.27-40
  • Descrição: Convolutional Neural Networks (CNNs) have emerged as a fundamental technology for machine learning. High performance and extreme energy efficiency are critical for deployments of CNNs, especially in mobile platforms such as autonomous vehicles, cameras, and electronic personal assistants. This paper introduces the Sparse CNN (SCNN) accelerator architecture, which improves performance and energy efficiency by exploiting the zero-valued weights that stem from network pruning during training and zero-valued activations that arise from the common ReLU operator. Specifically, SCNN employs a novel dataflow that enables maintaining the sparse weights and activations in a compressed encoding, which eliminates unnecessary data transfers and reduces storage requirements. Furthermore, the SCNN dataflow facilitates efficient delivery of those weights and activations to a multiplier array, where they are extensively reused; product accumulation is performed in a novel accumulator array. On contemporary neural networks, SCNN can improve both performance and energy by a factor of 2.7× and 2.3×, respectively, over a comparably provisioned dense CNN accelerator.
  • Editor: ACM
  • Idioma: Inglês

Buscando em bases de dados remotas. Favor aguardar.