In this video, we'll learn how to build Large Language Model (LLM) + Retrieval Augmented Generation (RAG) pipelines using open-source models from Hugging Face deployed on AWS SageMaker. We use the MiniLM sentence transformer to power our semantic search component with Pinecone.
📌 Code:
github.com/pinecone-io/exampl...
📕 Article:
www.pinecone.io/learn/sagemak...
🌲 Subscribe for Latest Articles and Videos:
www.pinecone.io/newsletter-si...
👋🏼 AI Consulting:
aurelio.ai
👾 Discord:
/ discord
Twitter: / jamescalam
LinkedIn: / jamescalam
00:00 Open Source LLMs on AWS SageMaker
00:27 Open Source RAG Pipeline
04:25 Deploying Hugging Face LLM on SageMaker
08:33 LLM Responses with Context
10:39 Why Retrieval Augmented Generation
11:50 Deploying our MiniLM Embedding Model
14:34 Creating the Context Embeddings
19:49 Downloading the SageMaker FAQs Dataset
20:23 Creating the Pinecone Vector Index
24:51 Making Queries in Pinecone
25:58 Implementing Retrieval Augmented Generation
30:00 Deleting our Running Instances
#artificialintelligence #nlp #aws #opensource #chatbot
Негізгі бет Ғылым және технология Hugging Face LLMs with SageMaker + RAG with Pinecone
Пікірлер: 30