Главная
Study mode:
on
1
- Intro
2
- MLPs vs KAN
3
- simple polynomials as Complex polynomials
4
- Toy Problem with KAN
5
- B-Splines
6
- KAN implementation Residual Activation Function
7
- Spline Grids and grid extension
8
- Fine grained training
9
- computational complexity
10
- Interpretability
11
- Experiments and results
12
- continual learning and catastrophic forgetting
13
- Choosing between KAN and MLP
Description:
Dive into a 23-minute video exploring Kolmogorov-Arnold Networks (KAN), a potential alternative to traditional Multi-Layer Perceptrons (MLPs) that power modern AI systems like ChatGPT, LLAMA, and DALLE. Learn about the mathematical foundations of KAN architectures, including B-splines and complex polynomials, through detailed explanations and practical implementations. Explore key concepts like residual activation functions, spline grids, and fine-grained training approaches while understanding the computational complexity and interpretability aspects of KAN. Compare experimental results between KAN and MLP networks, examining their performance in continual learning scenarios and addressing catastrophic forgetting. Gain insights into choosing between KAN and MLP architectures for different applications, supported by demonstrations using a toy problem and comprehensive implementation examples from the official PyKAN repository.

Kolmogorov-Arnold Networks: Understanding KAN Architecture and Comparison with MLPs

AI Bites
Add to list
0:00 / 0:00