Главная
Study mode:
on
1
Intro
2
What is a structured matrix
3
Why arent structured matrices more widespread
4
A large NLP application
5
Vanilla Chronic Product
6
Training Curves
7
Training Techniques
8
Training Results
9
Output Feature Vectors
10
Cometrics Adaptation
11
Regularization
12
Cometric Synchronization
13
Results
14
Training data
15
Controlling sparsity
16
Compression
17
Limitations
18
Conclusion
19
Rule of Thumb
20
More than 2 matrices
21
Why the perplexity increased
22
Posttraining decomposition
23
Sponsors
Description:
Explore a technique for extreme compression of LSTM models using sparse structured additive matrices in this tinyML Talks webcast. Learn about structured matrices derived from Kronecker products and their effectiveness in compressing neural networks. Discover the concept of "doping" - adding an extremely sparse matrix to a structured matrix - and how it provides additional degrees of freedom for parameters. Understand the challenges of training LSTMs with doped structured matrices, including co-matrix adaptation and the need for co-matrix dropout regularization. Examine empirical evidence demonstrating the applicability of these concepts to multiple structured matrices. Delve into state-of-the-art accuracy results at large compression factors across natural language processing applications. Compare the doped Kronecker product compression technique to previous compression methods, pruning, and low-rank alternatives. Investigate the deployment of doped KP on commodity hardware and the resulting inference run-time speed-ups. Cover topics such as training curves, output feature vectors, regularization techniques, controlling sparsity, and limitations of the approach. Read more

A Technique for Extreme Compression of LSTM Models

tinyML
Add to list