@sohnki

Memory Requirements for Convolutional Neural Network Hardware Accelerators

, , , and . 2018 IEEE International Symposium on Workload Characterization (IISWC), page 111-121. (September 2018)
DOI: 10.1109/IISWC.2018.8573527

Abstract

The rapid pace and successful application of machine learning research and development has seen widespread deployment of deep convolutional neural networks (CNNs). Alongside these algorithmic efforts, the compute- and memory-intensive nature of CNNs has stimulated a large amount of work in the field of hardware acceleration for these networks. In this paper, we profile the memory requirements of CNNs in terms of both on-chip memory size and off-chip memory bandwidth, in order to understand the impact of the memory system on accelerator design. We show that there are fundamental tradeoffs between performance, bandwidth, and on-chip memory. Further, this paper explores how the wide variety of CNNs for different application domains each have fundamentally different characteristics. We show that bandwidth and memory requirements for different networks, and occasionally for different layers within a network, can each vary by multiple orders of magnitude. This makes designing fast and efficient hardware for all CNN applications difficult. To remedy this, we outline heuristic design points that attempt to optimize for select dataflow scenarios.

Description

Memory Requirements for Convolutional Neural Network Hardware Accelerators - IEEE Conference Publication

Links and resources

Tags

community

  • @sohnki
  • @dblp
@sohnki's tags highlighted