Главная
Study mode:
on
1
Introduction
2
Outline
3
Deep Learning
4
Problems with Deep Learning
5
Data Centers and Deep Learning
6
Layers
7
Data Sets
8
Questions
9
Architectures
10
Number of operations
11
Convolutional layers
12
Comparison of architectures
13
Comparing architectures
14
Retraining feature maps
15
Quantizing parameters
16
Quantization experiment
17
Flops rate
18
Compensation during training
19
Quantization during training
20
Quantization for precision
21
Results
22
Shift Attention Layers
23
Clustering weights
24
Energy consumption
25
Summary
26
Conclusion
27
Questions and Recap
Description:
Explore a comprehensive review of compression methods for deep convolutional neural networks in this 58-minute tinyML Talks webcast. Delve into Professor Vincent Gripon's expertise as he discusses various techniques to compress and accelerate CNNs, including pruning, distillation, clustering, and quantization. Gain insights into the pros and cons of each method, understanding how to reduce the size of CNN architectures for tinyML devices without compromising accuracy. Learn about deep learning challenges, data centers, layers, datasets, architectures, and the number of operations in convolutional layers. Examine quantization experiments, flops rates, and energy consumption. Participate in a Q&A session to further explore the topic and clarify any questions about deploying CNNs on resource-constrained devices.

A Review of Compression Methods for Deep Convolutional Neural Networks

tinyML
Add to list