πŸ“š

RAG Full-Stack Development

This course is for developers with some AI and programming background, focusing on the principles, system composition, security considerations, and implementation of Retrieval-Augmented Generation (RAG). You will systematically learn the working mechanism of LLM and RAG, and master the complete E2E project construction method, including key modules such as index construction, query optimization, and permission control. The course adopts a combination of theory and practice to help you develop the ability to independently develop and deploy enterprise-level intelligent question-answering systems.

Day 1: RAG Theory and Core Process

Systematic Understanding from Concept to Technical Architecture

πŸ”

πŸ“š Today's Content

+
β€’

1.1 RAG Overview

What is RAG? (What)

Why is RAG needed? (Why)

Analysis of typical application scenarios

Introduce the overall technical architecture of RAG in combination with LangChain/LangGraph

β€’

1.2 Two-Stage Workflow of RAG

Overview of the core working mechanism of the RAG system

Analysis of the two-stage collaborative mode

πŸ“₯ Indexing Stage
β€’ Document loading and chunking - LangChain / LCEL's loading and chunking capabilities
β€’ Vectorization and Embedding models - selection of mainstream open source and commercial embedding models
β€’ Vector database access - integration of vector storage solutions such as pgvector, FAISS, etc.
πŸ”Ž Retrieving Stage
β€’ Retrieval technology principles: vector retrieval, similarity ranking
β€’ Prompt template design and integration of upstream and downstream links
β€’ Advanced retrieval strategies: Hybrid Search, Reranking
β€’ Retrieval link construction in LangChain/LangGraph
β€’

1.4 Tool Practice Demonstration

Use LangChain/LCEL to demonstrate the complete RAG process: document loading β†’ Embedding β†’ retrieval

Operate a vector database (taking pgvector as an example)

πŸ’‘ Learning Objectives

+

πŸ“¦ Appendix

+