TY - JOUR
T1 - IMPULSE
T2 - A 65-nm Digital Compute-in-Memory Macro with Fused Weights and Membrane Potential for Spike-Based Sequential Learning Tasks
AU - Agrawal, Amogh
AU - Ali, Mustafa
AU - Koo, Minsuk
AU - Rathi, Nitin
AU - Jaiswal, Akhilesh
AU - Roy, Kaushik
N1 - Publisher Copyright:
© 2018 IEEE.
PY - 2021
Y1 - 2021
N2 - The inherent dynamics of the neuron membrane potential in spiking neural networks (SNNs) allows the processing of sequential learning tasks, avoiding the complexity of recurrent neural networks. The highly sparse spike-based computations in such spatiotemporal data can be leveraged for energy efficiency. However, the membrane potential incurs additional memory access bottlenecks in current SNN hardware. To that effect, we propose a 10T-SRAM compute-in-memory (CIM) macro, specifically designed for state-of-the-art SNN inference. It consists of a fused weight ( WMEM) and membrane potential (VMEM) memory and inherently exploits sparsity in input spikes leading to 97.4% reduction in energy-delay product (EDP) at 85% sparsity (typical of SNNs considered in this work) compared to the case of no sparsity. We propose staggered data mapping and reconfigurable peripherals for handling different bit precision requirements of WMEM and VMEM , while supporting multiple neuron functionalities. The proposed macro was fabricated in 65-nm CMOS technology, achieving energy efficiency of 0.99 TOPS/W at 0.85-V supply and 200-MHz frequency for signed 11-bit operations. We evaluate the SNN for sentiment classification from the IMDB dataset of movie reviews and achieve within 1% accuracy difference and ∼ 5 × higher energy efficiency compared to a corresponding long short-term memory network.
AB - The inherent dynamics of the neuron membrane potential in spiking neural networks (SNNs) allows the processing of sequential learning tasks, avoiding the complexity of recurrent neural networks. The highly sparse spike-based computations in such spatiotemporal data can be leveraged for energy efficiency. However, the membrane potential incurs additional memory access bottlenecks in current SNN hardware. To that effect, we propose a 10T-SRAM compute-in-memory (CIM) macro, specifically designed for state-of-the-art SNN inference. It consists of a fused weight ( WMEM) and membrane potential (VMEM) memory and inherently exploits sparsity in input spikes leading to 97.4% reduction in energy-delay product (EDP) at 85% sparsity (typical of SNNs considered in this work) compared to the case of no sparsity. We propose staggered data mapping and reconfigurable peripherals for handling different bit precision requirements of WMEM and VMEM , while supporting multiple neuron functionalities. The proposed macro was fabricated in 65-nm CMOS technology, achieving energy efficiency of 0.99 TOPS/W at 0.85-V supply and 200-MHz frequency for signed 11-bit operations. We evaluate the SNN for sentiment classification from the IMDB dataset of movie reviews and achieve within 1% accuracy difference and ∼ 5 × higher energy efficiency compared to a corresponding long short-term memory network.
KW - Compute-in-memory (CIM)
KW - neuromorphic computing
KW - sentiment analysis
KW - spiking neural network (SNN)
KW - SRAM
UR - http://www.scopus.com/inward/record.url?scp=85111095229&partnerID=8YFLogxK
U2 - 10.1109/LSSC.2021.3092727
DO - 10.1109/LSSC.2021.3092727
M3 - Article
AN - SCOPUS:85111095229
SN - 2573-9603
VL - 4
SP - 137
EP - 140
JO - IEEE Solid-State Circuits Letters
JF - IEEE Solid-State Circuits Letters
M1 - 9466245
ER -