Главная
Study mode:
on
1
Intro
2
Why Called BitNet 1.58
3
Why Should I Care?
4
Math
5
Quantization Without BitNet
6
BitLinear Layer
7
What About Backpropagation?
8
How Many Gainz?
9
Bessie the BitNet
10
Testing the Base Model
11
Fine_Tuning for QA/Instructions
12
The Code
13
Diving into the Quantization
14
Good News and Bad News
15
What’s Next?
16
Takeaways
Description:
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only! Grab it Learn about the groundbreaking technique of 1-bit Large Language Models in this 47-minute technical video that explores the research paper "The Era of 1-bit LLMs." Discover how weights can be represented using only 0, 1, or -1 integers instead of traditional floating-point numbers. Explore the mathematical foundations, understand BitLinear layers, and delve into the implications for backpropagation. Follow along as the presenter demonstrates practical implementations, including base model testing and fine-tuning for question-answering tasks. Examine detailed code examples of quantization techniques, evaluate the advantages and limitations of this approach, and understand potential future developments in the field. Perfect for AI researchers, developers, and enthusiasts interested in model optimization and efficient machine learning architectures.

Understanding 1-Bit Large Language Models - From Theory to Implementation

Oxen
Add to list