Главная
Study mode:
on
1
Introduction
2
Outline
3
Agenda
4
Rooney Quantization
5
Original Weight
6
Scaling Factor
7
Convolution
8
Posttraining quantization
9
Quantization granularity
10
Perchannel quantization
11
Skating Factor
12
Clipping clipping clipping
13
Selecting clipping range
14
Fine tuning
15
Weight
16
Activation
17
Quantization Example
18
Quantization Notation
19
Quantization Results
20
Binary ternary quantization
21
Stochastic binarization
Description:
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only! Grab it Dive into the second part of a comprehensive lecture on quantization in machine learning, delivered by Prof. Song Han as part of MIT's 6.5940 course for Fall 2023. Explore advanced concepts such as Rooney Quantization, scaling factors, convolution techniques, and post-training quantization methods. Learn about quantization granularity, per-channel quantization, and the importance of clipping in the quantization process. Discover how to select optimal clipping ranges and fine-tune quantized models. Examine weight and activation quantization through practical examples, and gain insights into binary and ternary quantization techniques, including stochastic binarization. Access accompanying slides at efficientml.ai for a deeper understanding of these cutting-edge quantization strategies in efficient machine learning.

EfficientML.ai: Quantization Part II - Lecture 6

MIT HAN Lab
Add to list