Главная
Study mode:
on
1
Introduction
2
Computing system design
3
Transformer architecture
4
Uniform quantization
5
Uniform quantization scheme
6
Uniform continuation limits
7
Is it still useful
8
BCQ
9
Example
10
Critical problems
11
Lookup table
12
Transformer structure
13
Quantizing embedding layers
14
Mixed precision quantization
15
Encoder and Decoder
16
Retraining
17
Quantitation Results
18
Latency Improvements
19
Quantization
20
Q A
21
Strategic Partners
Description:
Explore extremely low-bit quantization techniques for Transformers in this tinyML Asia 2021 conference talk. Delve into the challenges of deploying Transformer architecture on resource-limited devices and learn about effective quantization strategies. Discover how different Transformer blocks contribute to model accuracy and inference computations, and understand the varying impacts of individual words within embedding blocks. Examine a proposed mixed precision quantization approach for representing Transformer weights using fewer than 3 bits, including a method for assigning different quantization bits to each word in an embedding block based on statistical properties. Gain insights into a novel matrix multiplication kernel that eliminates the need for dequantization steps. Cover topics such as computing system design, uniform quantization schemes, critical problems in quantization, and the Transformer structure. Explore quantization results, latency improvements, and participate in a Q&A session to deepen your understanding of this cutting-edge approach to optimizing Transformer models for mobile and edge devices. Read more

Extremely Low-Bit Quantization for Transformers - tinyML Asia 2021

tinyML
Add to list
0:00 / 0:00