Back to Browse

RAG Optimization: A Practical Overview for Improving Retrieval Augmented Generation

7.7K views
Jun 5, 2024
21:41

Optimizing retrieval augmented generation makes large language models more powerful and reliable, but off-the-shelf components yield lackluster results. Snorkel AI Principal Research Scientist Chris Glaze explains how to fine-tune multiple parts of rag systems—from document chunking to embedding models and to data enrichment—to ensure that LLM systems use their model's context window as effectively as possible. See more videos about RAG here: https://www.youtube.com/playlist?list=PLZePYakcDhmiPg-5JGfu20RaFV9RT_PQl Timestamps: 00:00 Introduction to RAG Optimization 01:36 Importance of Retrieval in RAG 02:38 Document Chunking Process 03:33 Techniques for Chunking Documents 08:32 Metadata Extraction and Its Value 10:02 Approaches to Information Extraction 11:06 Overview of Embeddings and Retrieval Models 12:20 Fine-Tuning Embeddings for Retrieval 13:14 Baseline Evaluation of Embedding Models 14:29 Data Development for Fine-Tuning 17:18 Creating a Training Set 18:40 Utilizing Relevant Scores 20:44 Summary of RAG Pipeline Optimization #enterpriseai #rag #machinelearning

Download

1 formats

Video Formats

360pmp428.3 MB

Right-click 'Download' and select 'Save Link As' if the file opens in a new tab.

RAG Optimization: A Practical Overview for Improving Retrieval Augmented Generation | NatokHD