Back to Browse

Reinforcement Learning From Human Feedback (RLHF) | Direct Preference Optimization (DPO) | Explained

11 views
Apr 25, 2026
18:32

📘 Notes: https://robosathi.com/docs/natural_language_processing/llm/ 🎥 NLP Playlist: https://www.youtube.com/playlist?list=PLnpa6KP2ZQxcDlHCeNiKbRhLWKVunQaxn 🎥 LLM: https://youtu.be/vEqaew-D28U 🎥 SFT: https://youtu.be/NTS0CuMItDY ✅ This video describes how RLHF helps us make align LLM outputs to human values, making it more safe and helpful. ✅ Here we will also understand the Direct Preference Optimization (DPO) technique used for RLHF in depth. 🕔 Time Stamp 🕘 00:00:00 - 00:01:02 Introduction 00:01:03 - 00:04:20 LLM Training Phases 00:04:21 - 00:07:12 Limitations of SFT 00:07:13 - 00:08:56 Reinforcement Learning From Human Feedback (RLHF) 00:08:57 - 00:14:59 Direct Preference Optimization (DPO) 00:15:00 - 00:17:34 Key Use Cases of RLHF 00:17:35 - 00:18:33 Next: BERT

Download

1 formats

Video Formats

360pmp420.7 MB

Right-click 'Download' and select 'Save Link As' if the file opens in a new tab.

Reinforcement Learning From Human Feedback (RLHF) | Direct Preference Optimization (DPO) | Explained | NatokHD