Главная
Study mode:
on
1
[] Yujian's preferred coffee
2
[] Takeaways
3
[] Please like, share, and subscribe to our MLOps channels!
4
[] The hero of the LLM space
5
[] Embeddings into Vector databases
6
[] What is large and what is small LLM consensus
7
[] QA Bot behind the scenes
8
[] Fun fact getting more context
9
[] RAGs eliminate the ability of LLMs to hallucinate
10
[] Critical part of the rag stack
11
[] Building citations
12
[] Difference between context and relevance
13
[] Missing prompt tooling
14
[] Similarity search
15
[] RAG Optimization
16
[] Interacting with LLMs and tradeoffs
17
[] RAGs not suited for
18
[] Fashion App
19
[] Multimodel Rags vs LLM RAGs
20
[] Multimodel use cases
21
[] Video citations
22
[] Wrap up
Description:
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only! Grab it Explore the complexities of Retrieval Augmented Generation (RAG) in this 49-minute MLOps podcast episode featuring Yujian Tang, Developer Advocate at Zilliz. Delve into the nuanced challenges developers face when implementing RAG, moving beyond industry oversimplifications. Learn about embedding vector databases, the consensus on large and small language models, and the intricacies of QA bots. Discover critical components of the RAG stack, including citation building, context vs. relevance, and similarity search. Examine RAG optimization techniques, discuss scenarios where RAG may not be suitable, and explore multimodal RAG applications. Gain insights into fashion app development and video citation methods while understanding the trade-offs in LLM interactions.

RAG Has Been Oversimplified - Exploring Complexities in Retrieval Augmented Generation

MLOps.community
Add to list
0:00 / 0:00