Back to Browse

L24: Next sentence prediction | BERT training objectives explained

3.1K views
Dec 26, 2024
6:05

Welcome to Lecture 24 of the course "Large Language Models" by Prof. Mitesh M.Khapra. Full Course: https://study.iitm.ac.in/ds/course_pages/BSCS5001.html Video Overview In this lecture, we break down the original BERT model’s training objectives: Masked Language Modeling (MLM) and Next Sentence Prediction (NSP). You’ll learn how MLM helps BERT predict masked words within a sentence and how NSP was designed to capture relationships between consecutive sentences. We will also explore how segment embeddings and positional embeddings are learned, and the role they play in enabling downstream NLP tasks. Finally, we’ll discuss why NSP was later dropped in many subsequent models, and what this shift meant for modern transformer architectures. About IIT Madras' online Bachelor of Science programme IIT Madras offers four-year BS programmes that aim to provide quality education to all, irrespective of age, educational background, or location. The BS programme has multiple levels, which provide flexibility to students to exit at any of these levels. Depending on the courses completed and credits earned, the learner can receive a Foundation Certificate from IITM CODE (Centre for Outreach and Digital Education), Diploma(s) from IIT Madras, or BSc/BS Degrees from IIT Madras. For more details, Visit: https://www.iitm.ac.in/academics/study-at-iitm/non-campus-bs-programmes #BERT #NLP #MachineLearning #DeepLearning #Transformers #MaskedLanguageModeling #NextSentencePrediction #LanguageModels #AI #ArtificialIntelligence #Embeddings #SegmentEmbeddings #PositionalEmbeddings #CLS #Tokenization #AIEducation #NeuralNetworks #LanguageRepresentation

Download

1 formats

Video Formats

360pmp412.7 MB

Right-click 'Download' and select 'Save Link As' if the file opens in a new tab.

L24: Next sentence prediction | BERT training objectives explained | NatokHD