A lecture that discusses gradient descent and stochastic gradient descent in deep neural networks. We discuss the effects of learning rates that are too large or too small. We discuss convergence rates for gradient descent and stochastic gradient descent for convex functions.
This lecture is from Northeastern University's CS 7150 Summer 2020 class on Deep Learning, taught by Paul Hand.
The notes are available at: http://khoury.northeastern.edu/home/hand/teaching/cs7150-summer-2020/Gradient_Descent_and_Stochastic_Gradient_Descent.pdf
References:
Li et al. 2018:
Li, Hao, Zheng Xu, Gavin Taylor, Christoph Studer, and Tom Goldstein. "Visualizing the loss landscape of neural nets." In Advances in Neural Information Processing Systems, pp. 6389-6399. 2018.
Download
0 formats
No download links available.
Gradient Descent and Stochastic Gradient Descent | NatokHD