Our most popular RAG solution for Azure has now been deployed thousands of times by developers using it across myriad domains, like meeting transcripts, research papers, HR documents, and industry manuals. Based on feedback from the community (and often, thanks to pull requests from the community!), we've added the most hotly requested features: support for multiple document types, chat history with Cosmos DB, user account and login, data access control, multimodal media ingestion, private deployment, and more.
This open-source RAG solution is powerful, but it can be intimidating to dive into the code yourself, especially now that it has so many optional features. That's why we're putting on a 10-part live series in January/February, diving deep into the solution and showing you all the ways you can use it.
Register for the whole series on Reactor or scroll down to learn about each session and register for individual sessions. We look forward to seeing you in the live chat and hearing how you're using the RAG solution for your own domain. See you in the streams! 👋🏻
The RAG solution for Azure
13 January, 2025 | 11:30 PM UTC | 3:30 PM PT
Register for the stream on Reactor
In this series, we'll dive deep into our most popular, fully-featured, and open-source RAG solution: https://aka.ms/ragchat Join us for the kick-off session, where we'll do a live demo of the RAG solution and explain how it all works. We'll step through the RAG flow from Azure AI Search to Azure OpenAI, deploy the app to Azure, and discuss the Azure architecture.
Customizing our RAG solution
15 January, 2025 | 11:30 PM UTC | 3:30 PM PT
Register for the stream on Reactor
In this series, we'll dive deep into our most popular, fully-featured, and open-source RAG solution: https://aka.ms/ragchat In our second session, we'll show you how to customize the RAG solution for your own domain - adding your own data, modifying the prompts, and personalizing the UI. Plus, we'll give you tips for local development for faster feature iteration.
Optimal retrieval with Azure AI Search
20 January, 2025 | 11:30 PM UTC | 3:30 PM PT
Register for the stream on Reactor
In this series, we dive deep into our most popular, fully-featured, and open-source RAG solution: https://aka.ms/ragchat Our RAG solution uses Azure AI Search to find matching documents, using state-of-the-art retrieval mechanisms. We'll dive into the mechanics of vector embeddings, hybrid search with RRF, and semantic ranking. We'll also discuss the data ingestion process, highlighting the differences between manual ingestion and integrated vectorization
Multimedia data ingestion
22 January, 2025 | 11:30 PM UTC | 3:30 PM PT
Register for the stream on Reactor
In this series, we dive deep into our most popular, fully-featured, and open-source RAG solution: https://aka.ms/ragchat Do your documents contain images or charts? Our RAG solution has two different approaches to handling multimedia documents, and we'll dive into both approaches in this session. The first approach is purely during ingestion time, where it replaces media in the documents with LLM-generated descriptions. The second approach stores images of the media alongside vector embeddings of the images, and sends both text and images to a multimodal LLM for question answering. Learn about both approaches in this session so that you can decide what to use for your app.
User login and data access control
27 January, 2025 | 11:30 PM UTC | 3:30 PM PT
Register for the stream on Reactor
In this series, we dive deep into our most popular, fully-featured, and open-source RAG solution: https://aka.ms/ragchat In our RAG flow, the app first searches a knowledge base for relevant matches to a user's query, then sends the results to the LLM along with the original question. What if you have documents that should only be accessed by a subset of your users, like a group or a single user? Then you need data access controls to ensure that document visibility is respected during the RAG flow. In this session, we'll show an approach using Azure AI Search with data access controls to only search the documents that can be seen by the logged in user. We'll also demonstrate a feature for user-uploaded documents that uses data access controls along with Azure Data Lake Storage Gen2.
Storing chat history
29 January, 2025 | 11:30 PM UTC | 3:30 PM PT
Register for the stream on Reactor
In this series, we dive deep into our most popular, fully-featured, and open-source RAG solution: https://aka.ms/ragchat Learn how we store chat history using either IndexedDB for client-side storage or Azure Cosmos DB for persistent storage. We'll discuss the API architecture and data schema choices, doing both a live demo of the app and a walkthrough of the code.
Adding speech input and output
3 February, 2025 | 11:30 PM UTC | 3:30 PM PT
Register for the stream on Reactor
In this series, we dive deep into our most popular, fully-featured, and open-source RAG solution: https://aka.ms/ragchat Our RAG solution includes optional features for speech input and output, powered either by the free browser SDKs or by the powerful Azure Speech API. We also offer a tight integration with the VoiceRAG solution, for those of you who want a real-time voice interface. Learn about all the ways you can add speech to your RAG chat in this session!
Private deployment
5 February, 2025 | 11:30 PM UTC | 3:30 PM PT
Register for the stream on Reactor
In this series, we dive deep into our most popular, fully-featured, and open-source RAG solution: https://aka.ms/ragchat To ensure that the RAG app can only be accessed within your enterprise network, you can deploy it to an Azure virtual network with private endpoints for each Azure service used. In this session, we'll show how to deploy the app to a virtual network that includes AI Search, OpenAI, Document Intelligence, and Blob storage. Then we'll log in to the virtual network using Azure Bastion with a virtual machine to demonstrate that we can access the RAG app from inside the network, and only inside the network.
Evaluating RAG answer quality
10 February, 2025 | 11:30 PM UTC | 3:30 PM PT
Register for the stream on Reactor
In this series, we dive deep into our most popular, fully-featured, and open-source RAG solution: https://aka.ms/ragchat How can you be sure that the RAG chat app answers are accurate, clear, and well formatted? Evaluation! In this session, we'll show you how to generate synthetic data and run bulk evaluations on your RAG app, using the azure-ai-evaluation SDK. Learn about GPT metrics like groundedness and fluency, and custom metrics like citation matching. Plus, discover how you can run evaluations on CI/CD, to easily verify that new changes don't introduce quality regressions.
Monitoring and tracing LLM calls
12 February, 2025 | 11:30 PM UTC | 3:30 PM PT
Register for the stream on Reactor
In this series, we dive deep into our most popular, fully-featured, and open-source RAG solution: https://aka.ms/ragchat When your RAG app is in production, observability is crucial. You need to know about performance issues, runtime errors, and LLM-specific issues like Content Safety filter violations. In this session, learn how to use Azure Monitor along with OpenTelemetry SDKs to monitor the RAG application.