Training Deep Neural Networks

29 / 49

Momentum optimization does not care about what previous gradients were. So gradient descent is better to converge faster.

See Answer

Note - Having trouble with the assessment engine? Follow the steps listed here


No hints are availble for this assesment

Loading comments...