Главная
Study mode:
on
1
Introduction
2
Scaling
3
Framework
4
Naive heuristic
5
Outline
6
Edge of Cows
7
Recurrence Equation
8
Gaussian PreActivations
9
The Edge of Chaos
10
Experiments
11
Gaussian regulations
12
Assumption of edge of chaos
13
Preservation of variance
14
Solution
15
Summary
16
Constraints
17
Density
18
activation functions
19
numerical approximations
20
training experiments
21
training losses
22
conclusion
23
future work
24
questions
25
data patterns
26
impossibility results
27
Cons
28
Training Loss
Description:
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only! Grab it Explore the intricacies of Gaussian pre-activations in neural networks through this 45-minute conference talk by Pierre Wolinski at the Finnish Center for Artificial Intelligence. Delve into the construction of activation functions and initialization distributions that ensure Gaussian pre-activations throughout network depth, even in narrow neural networks. Examine the critical review of Edge of Chaos claims and discover a unified view on pre-activations propagation. Gain insights into information propagation in deep and narrow neural networks, comparing ReLU and tanh activation functions with Kaiming and Xavier initializations. Learn about the speaker's background in neural network pruning, Bayesian neural networks, and current research on information propagation during initialization and training.

Gaussian Pre-Activations in Neural Networks: Myth or Reality?

Finnish Center for Artificial Intelligence FCAI
Add to list