Training Deep Neural Networks

You are currently auditing this course.
12 / 49

With SELU activation function even a 100 layer deep neural network preserves roughly mean 0 and standard deviation 1 across all layers avoiding the exploding/vanishing gradients problem.

See Answer

No hints are availble for this assesment

Loading comments...