Back to Browse

Gradient Descent: The Algorithm That Trains Every AI Model

11 views
May 6, 2026
4:18

Gradient descent is the engine behind every AI model ever trained. From ChatGPT to Stable Diffusion to self-driving cars β€” they all learn using the same simple rule: follow the slope downhill. In this episode, we break down exactly how it works: partial derivatives, the update rule, the learning rate, and modern tricks like momentum, Adam, and stochastic gradient descent. 🎯 CHAPTERS 00:00 Why gradient descent is the engine of AI 0:45 The blindfolded hiker intuition 1:45 Partial derivatives and the gradient 3:15 The update rule: w ← w βˆ’ Ξ±βˆ‡L 4:30 The learning rate β€” Goldilocks territory 6:00 Watching gradient descent learn a line 7:30 When the surface isn't a nice bowl 9:00 Momentum and Adam: modern tricks 11:00 Stochastic gradient descent (SGD) 12:30 From 2 parameters to GPT-4 πŸ“Œ KEY CONCEPTS β€’ Partial derivatives and the gradient vector β€’ The update rule: new w = old w βˆ’ Ξ± Γ— βˆ‚L/βˆ‚w β€’ Learning rate: too small, too big, just right β€’ Local minima, saddle points, plateaus β€’ Momentum and Adam optimizer β€’ Stochastic gradient descent (mini-batches) πŸ“š PREREQUISITE β€’ Episode 1 β€” Linear Regression: youtu.be/... πŸ“š THE SERIES Episode 2 of 10. Next: Neural Networks β€” how stacking linear functions with one simple trick lets us learn anything. πŸ“± WANT TO PRACTICE THE MATH? NovaMaths β€” SAT & ACT Math prep with 749+ exercises: https://www.novamaths.app 🌍 French channel: β€ͺ@MathsAcademy27‬ ───────────────────── Channel hosted by Julien, certified math teacher with 30 years of classroom experience. #GradientDescent #Adam #SGD #MachineLearning #AIMath

Download

0 formats

No download links available.

Gradient Descent: The Algorithm That Trains Every AI Model | NatokHD