Quickly build and deploy LLMs with Retrieval Augmented Generation, featuring Intel Accelerators and Azure

Logo
Presented by

Akash Shankaran (Lead Software Architect, Intel), Ron Abellera (Linux and AI Infrastructure Global Black belt, Microsoft), and Juan Pablo Norena (Field Software Engineer, Canonical)

About this talk

RAG offers a new way to maximize the capabilities of large language models (LLMs) to produce more accurate, context-aware, and informative responses. Join Akash Shankaran (Intel), Ron Abellera (Microsoft), and Juan Pablo Norena (Canonical) for a tutorial on how RAG can enhance your LLMs. The session will explore how to optimize LLMs with RAG using Charmed OpenSearch, which can serve multiple services like data ingestion, model ingestion, vector database, retrieval and ranking, and LLM connector. We will also show the architecture of our RAG deployment in Microsoft® Azure Cloud. Notably, the vector search capabilities of RAG are enhanced by Intel AVX® Acceleration, delivering faster processing and high-throughput performance for the RAG workflow.
Related topics:

More from this channel

Upcoming talks (5)
On-demand talks (436)
Subscribers (171163)
Get the most in depth information about the Ubuntu technology and services from Canonical. Learn why Ubuntu is the preferred Linux platform and how Canonical can help you make the most out of your Ubuntu environment.