Design space exploration of in-memory computing DNN accelerators

Pouya Houshmand and Jiacong Sun , Marian Verhelst Hardware-efficient AI and ML

Research goals: In recent years, in-memory computing (IMC) has emerged as a promising alternative to PE-based accelerators, by performing the MAC operations near/in the memory cells directly. This allows to greatly reduce access overheads and enables massive parallelization opportunities, with potential orders of magnitude improvements in energy efficiency and throughput. However these architectures suffer from lack of dataflow flexibility and thus potentially low utilization. The research focuses on exploring the algorithm-hardware design space for in-memory computing DNN accelerators.

Gap in the SotA: Recent IMC designs published in the literature are focused on analog IMC (AIMC), where the computation is carried out in the analog domain. While this approach ensures extreme energy efficiencies and massive parallelization, the analog nature of the computation compromises the output accuracy and the rigid structure of the computation's dataflow limits the spatial mapping possibilities. Digital in memory computing (DIMC) is lately gaining more interest as a valid alternative, thanks to its noise-free computation and more flexible spatial mapping possibilities, trading off added flexibility and accurate computation for less energy efficiency. Research work is needed to explore the design space of DIMC designs versus AIMC ones.

Research project: The ZigZag design space exploration tool is expanded with models for AIMC and DIMC specific building blocks. This allows to assess the impact of these technologies and related hardware parameters at the system level. 

Get in touch
Pouya Houshmand
Phd student
Jiacong Sun
Phd student
Marian Verhelst
Academic staff
Fundamental architecture for AIMC (a) and DIMC (b) and their mapping paradigm
Fundamental architecture for AIMC (a) and DIMC (b) and their mapping paradigm

Publications about this research topic

L. Mei*, P. Houshmand*, V. Jain, S. Giraldo and M. Verhelst, "ZigZag: Enlarging Joint Architecture-Mapping Design Space Exploration for DNN Accelerators," in IEEE Transactions on Computers, vol. 70, no. 8, pp. 1160-1174, 1 Aug. 2021 (* the authors have contributed equally)

P. Houshmand et al., "Opportunities and Limitations of Emerging Analog in-Memory Compute DNN Architectures," 2020 IEEE International Electron Devices Meeting (IEDM), 2020, pp. 29.1.1-29.1.4.

P. Houshmand*, G.Sarda*,  et al., "DIANA: An End-to-End Hybrid DIgital and ANAlog Neural Network SoC for the Edge," in IEEE Journal of Solid-State Circuits, 2022 (* the authors have contributed equally)

 

J. Sun, P. Houshmand, M. Verhelst, "Analog or digital in-memory computing? Benchmarking through quantitative modeling", in 2023 IEEE/ACM International Conference on Computer-Aided Design (ICCAD), 2023

Other research topics in Hardware-efficient AI and ML

Anda: Unlocking Efficient LLM Inference with a Variable-Length Grouped Activation Data Format
Hardware-efficient AI and ML
Man Shi, Arne Symons, Robin Geens, and Chao Fang | Marian Verhelst
Massive parallelism for combinatorial optimisation problems
Hardware-efficient AI and ML
Toon Bettens and Sofie De Weer | Wim Dehaene and Marian Verhelst
Carbon-aware Design Space Exploration for AI Accelerators
Hardware-efficient AI and ML
Jiacong Sun | Georges Gielen and Marian Verhelst
Decoupled Control Flow and Memory Orchestration in the Vortex GPGPU
Hardware-efficient AI and ML
Giuseppe Sarda | Marian Verhelst
Automated Causal CNN Scheduling Optimizer for Real-Time Edge Accelerators
Hardware-efficient AI and ML
Jun Yin | Marian Verhelst
A Scalable Heterogenous Multi-accelerator Platform for AI and ML
Hardware-efficient AI and ML
Ryan Antonio | Marian Verhelst
Uncertainty-Aware Design Space Exploration for AI Accelerators
Hardware-efficient AI and ML
Jiacong Sun | Georges Gielen and Marian Verhelst
Integer GEMM Accelerator for SNAX
Hardware-efficient AI and ML
Xiaoling Yi | Marian Verhelst
Improving GPGPU micro architecture for future AI workloads
Hardware-efficient AI and ML
Giuseppe Sarda | Marian Verhelst
SRAM based digital in memory compute macro in 16nm
Hardware-efficient AI and ML
Weijie Jiang | Wim Dehaene
Scalable large array nanopore readouts for proteomics and next-generation sequencing
Analog and power management circuits, Hardware-efficient AI and ML, Biomedical circuits and sensor interfaces
Sander Crols | Filip Tavernier and Marian Verhelst
Multi-core architecture exploration for layer-fused deep learning acceleration
Hardware-efficient AI and ML
Arne Symons | Marian Verhelst
HW-algorithm co-design for Bayesian inference of probabilistic machine learning
Ultra-low power digital SoCs and memories, Hardware-efficient AI and ML
Shirui Zhao | Marian Verhelst
Design space exploration for machine learning acceleration
Hardware-efficient AI and ML
Arne Symons | Marian Verhelst
Enabling Fast Exploration of the Depth-first Scheduling Space for DNN Accelerators
Hardware-efficient AI and ML
Arne Symons | Marian Verhelst
Optimized deployment of AI algorithms on rapidly-changing heterogeneous multi-core compute platforms
Ultra-low power digital SoCs and memories, Hardware-efficient AI and ML
Josse Van Delm | Marian Verhelst
High-throughput high-efficiency SRAM for neural networks
Ultra-low power digital SoCs and memories, Hardware-efficient AI and ML
Wim Dehaene and Marian Verhelst
Heterogeneous Multi-core System-on-Chips for Ultra Low Power Machine Learning Application at the Edge
Hardware-efficient AI and ML
Pouya Houshmand, Giuseppe Sarda, and Ryan Antonio | Marian Verhelst

Want to work with us?

Get in touch or discover the way we can collaborate.