Главная
Study mode:
on
1
Open Source LLMs on AWS SageMaker
2
Open Source RAG Pipeline
3
Deploying Hugging Face LLM on SageMaker
4
LLM Responses with Context
5
Why Retrieval Augmented Generation
6
Deploying our MiniLM Embedding Model
7
Creating the Context Embeddings
8
Downloading the SageMaker FAQs Dataset
9
Creating the Pinecone Vector Index
10
Making Queries in Pinecone
11
Implementing Retrieval Augmented Generation
12
Deleting our Running Instances
Description:
Learn how to build Large Language Model (LLM) and Retrieval Augmented Generation (RAG) pipelines using open-source models from Hugging Face deployed on AWS SageMaker in this comprehensive video tutorial. Explore the implementation of semantic search using the MiniLM sentence transformer with Pinecone. Discover the process of deploying Hugging Face LLMs on SageMaker, generating LLM responses with context, and understanding the benefits of Retrieval Augmented Generation. Follow along as the instructor demonstrates deploying the MiniLM embedding model, creating context embeddings, and setting up a Pinecone vector index using the SageMaker FAQs dataset. Gain practical insights into making queries in Pinecone and implementing RAG for improved AI-powered applications. The tutorial also covers essential steps for managing and deleting running instances to optimize resource usage.

Hugging Face LLMs with SageMaker - RAG with Pinecone

James Briggs
Add to list
0:00 / 0:00