Главная
Study mode:
on
1
Introduction
2
Theory
3
Main idea
4
Implementation
5
SelfExtend LLM
6
Deep Dive
7
Smooth Transition
8
Benchmark Data
9
Publication
10
Code Implementation
Description:
Learn how to extend the context length of Large Language Models (LLMs) during inference through a technical deep dive video that introduces grouped self-attention as an alternative to classical transformer self-attention mechanisms. Explore the challenges of out-of-distribution issues related to positional encoding when LLMs process text sequences beyond their pre-training context window. Examine implementation details, smooth transition techniques, and benchmark data while following along with code demonstrations based on the research paper "LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuning." Master practical solutions for handling longer sequences in neural networks without requiring model retraining or fine-tuning.

Self-Extending LLM Context Windows Using Grouped Self-Attention

Discover AI
Add to list