Build and deploy LLMs | Retrieval Augmented Generation | Data & AI Masters | Intel | Microsoft Azure

Estimated read time 1 min read

Post Content

​ Retrieval-augmented generation (RAG) offers a new way to maximise the capabilities of large language models (LLMs) to produce more accurate, context-aware, and informative responses. Join Akash Shankaran (Intel), Ron Abellera (Microsoft), and Juan Pablo Norena (Canonical) for a tutorial on how RAG can enhance your LLMs.

We will also explore optimising LLMs with RAG using Charmed OpenSearch, which can serve multiple services like data ingestion, model ingestion, vector database, retrieval and ranking, and LLM connector. We will also show the architecture of our RAG deployment in Microsoft® Azure Cloud. In addition, the RAG’s vector search capabilities are using Intel AVX® Acceleration, which enables faster and high-throughput generation of the RAG process. Learn more at https://canonical.com/solutions/ai and https://canonical.com/data   Read More Canonical Ubuntu 

#linux

You May Also Like

More From Author