The saying ""a picture is worth a thousand words"" encapsulates the immense potential of visual data. But most retrieval-augmented generation (RAG) applications rely only on text. This session applies RAG to multimodal use cases. It focuses on embeddings and attributed question answering to retrieve data. We’ll begin with a high-level architecture and quickly dive into a practical demo. Attendees will learn to create powerful LLM-based workflows and embed them in existing applications.
Speakers: Shilpa Kancharla, Jeff Nelson
Resources:
Try Gemini in Vertex AI → goo.gle/3Vttolh
Watch more:
Check out all the AI videos at Google I/O 2024 → goo.gle/io24-a...
Check out all the Cloud videos at Google I/O 2024 → goo.gle/io24-c...
Subscribe to Google Developers → goo.gle/develo...
#GoogleIO
Event: Google I/O 2024
Негізгі бет How to build Multimodal Retrieval-Augmented Generation (RAG) with Gemini
Пікірлер: 53