Главная
Study mode:
on
1
- Intro
2
- Environment Setup
3
- Function review
4
- Source Document
5
- Starting the project
6
- parse_file
7
- Understanding embeddings
8
- Implementing embeddings
9
- Timing embedding
10
- Caching embeddings
11
- Prompt embedding
12
- Cosine similarity for embedding comparison
13
- Brainstorming improvements
14
- Giving context to our LLM
15
- CLI input
16
- Next steps
Description:
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only! Grab it Dive into a comprehensive 16-minute tutorial on implementing Retrieval Augmented Generation (RAG) from scratch using Python and Ollama. Learn how to parse and manipulate documents, explore the concept of embeddings for describing abstract ideas, and implement an effective method for surfacing relevant document sections based on queries. Follow along to build a script that enables a locally-hosted Language Model to interact with your own documents. Gain insights into environment setup, function implementation, embedding techniques, caching strategies, and cosine similarity for comparison. Explore potential improvements and discover how to provide context to your LLM. By the end, you'll have a solid foundation for creating RAG systems and enhancing LLM interactions with custom datasets.

RAG from the Ground Up with Python and Ollama - Building a Document Interaction System

Decoder
Add to list