What is RAG? The Complete Tutorial - From Scratch to Deployed API on Production | LangChain & Ollama
Автор: Venelin Valkov
Загружено: 2025-07-19
Просмотров: 1407
Ever wondered how to make an LLM an expert on YOUR private documents? The answer is Retrieval-Augmented Generation (RAG). While stuffing context works for small files, it's slow, expensive, and fails at scale. RAG is the industry-standard solution.
In this complete, step-by-step tutorial, you will learn the fundamentals of RAG by building a system from the ground up. We'll start with first principles using Python and Scikit-learn, refactor our system with LangChain, wrap it in a streaming FastAPI, and finally deploy it as a production-ready Docker container.
PDF from the video: https://cdn.prod.website-files.com/60...
AI Bootcamp: https://www.mlexpert.io/
LinkedIn: / venelin-valkov
Follow me on X: / venelin_valkov
Discord: / discord
Subscribe: http://bit.ly/venelin-subscribe
GitHub repository: https://github.com/curiousily/AI-Boot...
👍 Don't Forget to Like, Comment, and Subscribe for More Tutorials!
00:00 - What is RAG?
05:58 - Project setup and dependencies
07:04 - Build a retriever
10:52 - Simple RAG
13:25 - Chat with PDF file
15:31 - Tracing and observability with MLflow
19:00 - RAG Rest API with FastAPI
24:02 - Docker container and compose
25:56 - Deploy to production
28:36 - Conclusion
Join this channel to get access to the perks and support my work:
/ @venelin_valkov
Доступные форматы для скачивания:
Скачать видео mp4
-
Информация по загрузке: