Email: eyeriss at mit dot edu
We will be giving a two day short course on Designing Efficient Deep Learning Systems on July 17-18, 2023 on MIT Campus (with a virtual option). To find out more, please visit MIT Professional Education.
Updated link to our book on Efficient Processing of Deep Neural Networks at here.
New article on "How to Evaluate Deep Neural Network Processors: TOPS/W (Alone) Considered Harmful" in SSCS Magazine is now available here.
Our book on Efficient Processing of Deep Neural Networks is now available here.
Excerpt of forthcoming book on Efficient Processing of Deep Neural Networks, Chapter on "Key Metrics and Design Objectives" available here.
Videos of ISCA tutorial on Timeloop/Accelergy Tutorial: Tools for Evaluating Deep Neural Network Accelerator Designs available here.
Our book on Efficient Processing of Deep Neural Networks now available for pre-order here.
Excerpt of forthcoming book on Efficient Processing of Deep Neural Networks, Chapter on "Advanced Technologies" available here.
Video and slides of NeurIPS tutorial on Efficient Processing of Deep Neural Networks: from Algorithms to Hardware Architectures available here.
We will be giving a two day short course on Designing Efficient Deep Learning Systems at MIT in Cambridge, MA on July 20-21, 2020. To find out more, please visit MIT Professional Education.
Slides for ICIP tutorial on Efficient Image Processing with Deep Neural Networks available here.
Code released for NetAdapt: Platform-Aware Neural Network Adaptation for Mobile Applications here.
Deep neural networks (DNNs) are currently widely used for many AI applications including computer vision, speech recognition, robotics, etc. While DNNs deliver state-of-the-art accuracy on many AI tasks, it comes at the cost of high computational complexity. Accordingly, designing efficient hardware architectures for deep neural networks is an important step towards enabling the wide deployment of DNNs in AI systems.
This tutorial provides a brief recap on the basics of deep neural networks and is for those who are interested in understanding how those models are mapping to hardware architectures. We will provide frameworks for understanding the design space for deep neural network accelerators including managing data movement, handling sparsity, and importance of flexibility. This is an intermediate-level tutorial that will go beyond the material in the previous incarnations of this tutorial.
An overview paper based on the tutorial "Efficient Processing of Deep Neural Networks: A Tutorial and Survey" is available here.
Our book based on the tutorial "Efficient Processing of Deep Neural Networks" is available here.
This book provides a structured treatment of the key principles and techniques for enabling efficient processing of deep neural networks (DNNs). The book includes background on DNN processing; a description and taxonomy of hardware architectural approaches for designing DNN accelerators; key metrics for evaluating and comparing different designs; features of the DNN processing that are amenable to hardware/algorithm co-design to improve energy efficiency and throughput; and opportunities for applying new technologies. Readers will find a structured introduction to the field as well as a formalization and organization of key concepts from contemporary works that provides insights that may spark new ideas.
An excerpt of the book on "Key Metrics and Design Objectives" and "Advanced Technologies" available at here.
ISCA 2017, CICS/MTL 2017, MICRO 2016
@article{2017_dnn_piee,
title={Efficient processing of deep neural networks: A tutorial and survey},
author={Sze, Vivienne and Chen, Yu-Hsin and Yang, Tien-Ju and Emer, Joel},
journal={Proceedings of the IEEE},
year={2017}
}