Design space exploration for machine learning acceleration

Arne Symons , Marian Verhelst Hardware-efficient AI and ML

In this research topic we focus on finding the optimal solutions for accelerating the performance of machine learning algorithms. This involves exploring different design and mapping options and evaluating their effectiveness in terms of speed, accuracy, and energy consumption. The ultimate goal is to find the best trade-off between these factors to produce a fast and efficient machine learning accelerator that can be used in a variety of applications. The research in this area involves the use of advanced tools and techniques, including simulation, modeling, and optimization, to help guide the design process and ensure that the resulting accelerator is optimal for the specific use case.

To this extent, we develop ZigZag, an open-source hardware architecture-mapping design space exploration framework for deep learning accelerators. ZigZag relies on an analytical cost model capable of accurately characterizing hardware performance of single-core accelerators.

Using ZigZag we have built multiple extensions which explore multi-core accelerators, depth-first processing, novel NN architectures such as transformer networks, etc.

Get in touch
Arne Symons
Phd student
Marian Verhelst
Academic staff
Overview of the ZigZag framework. Stages are customizable for tailored cost model, mapping optimization, architecture exploration, etc.
Overview of the ZigZag framework. Stages are customizable for tailored cost model, mapping optimization, architecture exploration, etc.

Other research topics in Hardware-efficient AI and ML

Automated Causal CNN Scheduling Optimizer for Real-Time Edge Accelerators
Hardware-efficient AI and ML
Jun Yin | Marian Verhelst
A Scalable Heterogenous Multi-accelerator Platform for AI and ML
Hardware-efficient AI and ML
Ryan Antonio | Marian Verhelst
Uncertainty-Aware Design Space Exploration for AI Accelerators
Hardware-efficient AI and ML
Jiacong Sun | Georges Gielen and Marian Verhelst
Integer GEMM Accelerator for SNAX
Hardware-efficient AI and ML
Xiaoling Yi | Marian Verhelst
Improving GPGPU micro architecture for future AI workloads
Hardware-efficient AI and ML
Giuseppe Sarda | Marian Verhelst
SRAM based digital in memory compute macro in 16nm
Hardware-efficient AI and ML
Weijie Jiang | Wim Dehaene
Scalable large array nanopore readouts for proteomics and next-generation sequencing
Analog and power management circuits, Hardware-efficient AI and ML, Biomedical circuits and sensor interfaces
Sander Crols | Filip Tavernier and Marian Verhelst
Design space exploration of in-memory computing DNN accelerators
Hardware-efficient AI and ML
Pouya Houshmand and Jiacong Sun | Marian Verhelst
Multi-core architecture exploration for layer-fused deep learning acceleration
Hardware-efficient AI and ML
Pouya Houshmand and Arne Symons | Marian Verhelst
HW-algorithm co-design for Bayesian inference of probabilistic machine learning
Ultra-low power digital SoCs and memories, Hardware-efficient AI and ML
Shirui Zhao | Marian Verhelst
Enabling Fast Exploration of the Depth-first Scheduling Space for DNN Accelerators
Hardware-efficient AI and ML
Arne Symons | Marian Verhelst
Optimized deployment of AI algorithms on rapidly-changing heterogeneous multi-core compute platforms
Ultra-low power digital SoCs and memories, Hardware-efficient AI and ML
Josse Van Delm | Marian Verhelst
High-throughput high-efficiency SRAM for neural networks
Ultra-low power digital SoCs and memories, Hardware-efficient AI and ML
Wim Dehaene and Marian Verhelst
Heterogeneous Multi-core System-on-Chips for Ultra Low Power Machine Learning Application at the Edge
Hardware-efficient AI and ML
Pouya Houshmand, Giuseppe Sarda, and Ryan Antonio | Marian Verhelst

Want to work with us?

Get in touch or discover the way we can collaborate.