Large Language Models (LLMs) Basics
Gain an in-depth understanding of components like Transformer's self-attention mechanism and positional encoding, implement a simplified Transformer from scratch, master the differences between models like GPT and BERT, and understand the relationship between model scale and emergent capabilities, as well as the characteristics of mainstream open-source models (Llama, Mistral, etc.).
Prompt Engineering
Covers zero-shot/few-shot prompting, chain-of-thought, self-consistency, and the ReAct framework; emphasizes structured prompts (role setting, output specifications); introduces automatic prompt optimization techniques (APE, OPRO).
RAG Architecture
Breaks down RAG components: document splitting, embedding model selection, vector database selection, retrieval optimization, and re-ranking techniques; provides code examples to convert unstructured documents into knowledge bases and improve the accuracy of generated content.
AI Agents
Explains core concepts like tool use, planning, memory, and multi-agent collaboration; through cases, master building systems for autonomous task decomposition, API calling, and long-term memory maintenance, covering frameworks like LangChain Agents and AutoGPT.
Vector Databases
Compares the characteristics of mainstream databases like Milvus and Pinecone; explains embedding model selection and fine-tuning, vector indexing algorithms (HNSW, IVF), and hybrid retrieval strategies.