High-throughput high-efficiency SRAM for neural networks

, Wim Dehaene and Marian Verhelst Ultra-low power digital SoCs and memories Hardware-efficient AI and ML

SRAM is an important component to store the weight and activations in a neural network chip. And for the parallelism, especially when cooperating with a large-scale PE array, multiple words would be addressed from the memory in the same cycle. Considering the significant portion consumed by memory operations, the high SRAM bandwidth could favour low latency and high energy efficiency of the network. Our target, therefore, is to provide neural networks with a high bandwidth SRAM for multiple-port access and promising density, power as well as frequency.

SotA: Multi-port SRAMs provided by foundries get use of multi-port bitcells and individual peripherals for every port. Although such designs offer a high bandwidth, they are facing the large bitcell area, i.e., low density, and serious voltage/frequency constraints as multi-port bitcells are vulnerable to access failures. Banked single-port SRAM is another approach, yet its low addressing flexibility leads to degraded throughput and complex mapping strategy between computation parts and memory components.

Our approach: This research explores the hierarchical multi-port SRAM design that multiple addressing is implemented for small bitcell matrixes. Inside the bitcell matrix, high-density single-port bitcells are used cooperating with pitch-matched local peripherals. Then, different voltage swings are assigned to global and local signals accordingly to save energy while maintain the accessibility.

Recent result: The first prototype would be carried out with GlobalFoundries 12nm LP process in late 2023. The target is to enable 4-port access with expected efficiency.

 

Get in touch
Wim Dehaene
Academic staff
Marian Verhelst
Academic staff

Other research topics in Ultra-low power digital SoCs and memories and Hardware-efficient AI and ML

Automated Causal CNN Scheduling Optimizer for Real-Time Edge Accelerators
Hardware-efficient AI and ML
Jun Yin | Marian Verhelst
A Scalable Heterogenous Multi-accelerator Platform for AI and ML
Hardware-efficient AI and ML
Ryan Antonio | Marian Verhelst
Uncertainty-Aware Design Space Exploration for AI Accelerators
Hardware-efficient AI and ML
Jiacong Sun | Georges Gielen and Marian Verhelst
Activity-independent variability resilience for complex ultra-low voltage digital ICs
Ultra-low power digital SoCs and memories
Clara Nieto Taladriz Moreno | Wim Dehaene
Integer GEMM Accelerator for SNAX
Hardware-efficient AI and ML
Xiaoling Yi | Marian Verhelst
Improving GPGPU micro architecture for future AI workloads
Hardware-efficient AI and ML
Giuseppe Sarda | Marian Verhelst
SRAM based digital in memory compute macro in 16nm
Hardware-efficient AI and ML
Weijie Jiang | Wim Dehaene
Scalable large array nanopore readouts for proteomics and next-generation sequencing
Analog and power management circuits, Hardware-efficient AI and ML, Biomedical circuits and sensor interfaces
Sander Crols | Filip Tavernier and Marian Verhelst
Design space exploration of in-memory computing DNN accelerators
Hardware-efficient AI and ML
Pouya Houshmand and Jiacong Sun | Marian Verhelst
Multi-core architecture exploration for layer-fused deep learning acceleration
Hardware-efficient AI and ML
Pouya Houshmand and Arne Symons | Marian Verhelst
HW-algorithm co-design for Bayesian inference of probabilistic machine learning
Ultra-low power digital SoCs and memories, Hardware-efficient AI and ML
Shirui Zhao | Marian Verhelst
Design space exploration for machine learning acceleration
Hardware-efficient AI and ML
Arne Symons | Marian Verhelst
Enabling Fast Exploration of the Depth-first Scheduling Space for DNN Accelerators
Hardware-efficient AI and ML
Arne Symons | Marian Verhelst
Automated in-situ monitoring for variability resilient and energy efficient digital circuits
Ultra-low power digital SoCs and memories
Clara Nieto Taladriz Moreno | Wim Dehaene
Optimized deployment of AI algorithms on rapidly-changing heterogeneous multi-core compute platforms
Ultra-low power digital SoCs and memories, Hardware-efficient AI and ML
Josse Van Delm | Marian Verhelst
Ultrasound wave based body area networks
Analog and power management circuits, Ultra-low power digital SoCs and memories, Biomedical circuits and sensor interfaces
Wim Dehaene and Marian Verhelst
Heterogeneous Multi-core System-on-Chips for Ultra Low Power Machine Learning Application at the Edge
Hardware-efficient AI and ML
Pouya Houshmand, Giuseppe Sarda, and Ryan Antonio | Marian Verhelst
SRAM Macro Design
Ultra-low power digital SoCs and memories
Bob Vanhoof | Wim Dehaene

Want to work with us?

Get in touch or discover the way we can collaborate.