Главная
Study mode:
on
1
Introduction
2
Strategic Partners
3
Welcome
4
Agenda
5
Deep Neural Networks
6
Image Classification Networks
7
Hardware accelerators
8
Motivation Analysis
9
Model Pruning
10
Layer Fusion
11
Methodology
12
Results
13
Extended Design Space
14
Mobilenet V1
15
Conclusion
16
Questions
17
Multilayer Fusion
18
Crowd Sponsors
Description:
Explore a comprehensive methodology for unifying design space exploration and model compression in deep learning accelerators for TinyML applications. Delve into the challenges of deploying Deep Learning models on resource-constrained embedded devices and learn about SuperSlash, an innovative solution that combines Design Space Exploration (DSE) and Model Compression techniques. Discover how SuperSlash estimates off-chip memory access volume overhead, evaluates data reuse strategies, and implements layer fusion to optimize performance. Gain insights into the pruning process guided by a ranking function based on explored off-chip memory access costs. Examine the application of this technique to fit large DNN models on accelerators with limited computational resources, using examples such as MobileNet V1. Engage with a detailed analysis of the extended design space, multilayer fusion, and the impact of these strategies on TinyML implementations.

TinyML Talks Pakistan - SuperSlash - Unifying Design Space Exploration and Model Compression

tinyML
Add to list
00:00
00:00