Главная
Study mode:
on
1
intro
2
preamble
3
mechanistic interpretability
4
neural network representations
5
qualities of representations
6
decomposability
7
linearity
8
linear composition as a compression scheme
9
demands of linearity
10
the linear representation puzzle
11
neuron - feature requirements
12
experience with llms
13
the superposition hypothesis
14
sparsity
15
recovering features in superposition
16
demands of linearity
17
feature exploration
18
thanks
Description:
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only! Grab it Explore the concept of superposition in large language model feature representations in this 47-minute conference talk from Conf42 LLMs 2024. Delve into mechanistic interpretability, neural network representations, and the qualities of these representations. Examine decomposability and linearity in depth, including linear composition as a compression scheme and its demands. Investigate the linear representation puzzle and neuron-feature requirements before diving into the superposition hypothesis. Analyze sparsity and learn techniques for recovering features in superposition. Conclude with a discussion on feature exploration in large language models.

Superposition in LLM Feature Representations

Conf42
Add to list