tinyML Talks on November 16, 2021 “SuperSlash: Unifying Design Space Exploration and Model Compression methodology for design of deep learning accelerators for TinyML” by Rehan Hafiz

Announcing our next tinyML Talks webcast! Rehan Hafiz from Information Technology University will present SuperSlash: Unifying Design Space Exploration and Model Compression methodology for design of deep learning accelerators for TinyML on November 16, 2021 at 8:00 AM Pacific Time.

Forum November 16

IMPORTANT: Please register here

Deploying Deep Learning (DL) models on resource-constrained embedded devices is a challenging task. The limited on-chip memory on such devices results in increased off-chip memory access volume, thus limiting the size of DL models that can be efficiently realized in such systems. Sophisticated DSE (Design Space Exploration) schemes have been developed in the past to reduce the off-chip memory access volume. However, DSE alone cannot reduce the amount of off-chip memory accesses beyond a certain point due to the fixed model size. Model compression via pruning can be employed to reduce the size of the model and the associated off-chip memory accesses. However, we found that pruned models with even the same accuracy and model size may require a different number of off-chip memory accesses depending upon the pruning strategy adopted. Furthermore, the classical pruning schemes are not guided by the goals of DSE. In this talk we discuss SuperSlash, a unified solution for DSE and Model Compression. SuperSlash estimates off-chip memory access volume overhead of each layer of a deep learning model by exploring multiple design candidates. In particular, it evaluates multiple data reuse strategies for each layer, along with the possibility of layer fusion. Layer fusion aims at reducing the off-chip memory access volume by avoiding the intermediate off-chip storage of a layer’s output and directly using it for processing of the subsequent layer. SuperSlash then guides the pruning process via a ranking function, which ranks each layer according to its explored off-chip memory access cost. The talk shall thus present a technique to jointly perform the pruning and DSE to fit in large DNN models on accelerators with low computational resources.

Rehan Hafiz received his Ph.D. degree in Electrical Engineering from the University of Manchester, United Kingdom, in 2008. He is currently with Information Technology University (ITU), Lahore, as a Professor in the Faculty of Engineering. He founded and directed the Vision Processing Lab (VISpro) that focuses on areas like Vision System Design, Approximate Computing, Design of Application-Specific Hardware Accelerators, Deep Learning, FPGA based design, and applied image and video processing. Apart from several publications in these areas, he holds multiple patents in the US, South Korean, and Pakistan patent offices.

=========================

IMPORTANT: Please register here

Once registered, you will receive a link and dial in information to Zoom teleconference by email, that you can also add to your calendar.

We encourage you to register earlier since on-line broadcast capacity may be limited.

Note: tinyML Talks slides and videos will be available here on the tinyML Forum and the tinyML YouTube Channel afterwards for those who missed the live session. Please take a moment and subscribe to the YouTube channel today

Feel free to ask your questions on this thread and keep the conversation going!