Building Intelligent Apps with RAG on Kubernetes: From Raw Data to Real-Time Insights
Workshop Overview
Welcome! This lab will illustrate how to combine event-driven serverless applications, data science pipelines, and a vector database to solve a real-world enterprise challenge using Large Language Models (LLMs).
The scenario for this lab is that we all work for "Parasol Insurance", a large company looking to leverage AI to improve its IT support operations. Our goal is to build a prototype Retrieval-Augmented Generation (RAG) system that taps into the vast knowledge base of closed ServiceNow tickets, making it instantly searchable for support engineers through a conversational AI assistant.
AI Lab Assistant
Throughout this workshop, you have access to an AI-powered lab assistant that can help you navigate the material, answer questions, and troubleshoot issues. The assistant is integrated directly into the documentation and has knowledge of the workshop content, OpenShift operations, and the technologies covered in this lab.
The AI assistant can help you to:
-
Ask questions about the technology - Learn more about RAG, vector databases, data science pipelines, or any other technology covered in the workshop
-
Get troubleshooting advice - When things go wrong, describe what you’re experiencing. Be specific: "In exercise 3, I deployed the ServiceNow app but I’m getting a 404 error when trying to access it" or "What is wrong with my milvus deployment in my project?"
-
Clarify workshop instructions - If any step in the workshop is unclear, ask for clarification or additional context
-
Understand OpenShift resources - Get help with pods, deployments, services, routes, and other Kubernetes/OpenShift concepts
To get the best results, be explicit and specific in your questions. Include details about which exercise you’re working on, what you’ve tried, and what error messages or unexpected behavior you’re seeing.
Disclaimer
This lab is an example of what a developer could build using Red Hat OpenShift AI. Red Hat OpenShift AI itself does not include a ServiceNow integration or a vector database out-of-the-box.
This lab makes use of a Large Language Model (LLM), the docling document parsing library, and the Milvus vector database. These models and third-party tools are not included in the Red Hat OpenShift AI product. They are provided as a convenience for this lab and as a learning experience.
| The quality of the models and the architecture are suitable for a prototype. Choosing the right components for a production environment is a complex task that requires significant experimentation and tuning. This lab does not cover production-level architecture design. |
Timetable
This is a tentative timetable for the materials that will be presented.
| Name | Duration | Type | Description |
|---|---|---|---|
The Parasol Company Scenario |
15 |
Presentation |
|
Building the API-to-RAG Pipeline |
45 |
Hands-On |
|
Event-Driven PDF Pipeline |
20 |
Hands-On |
|
Contributing
If you are interested in contributing to this project, consult this GitHub Repo: https://github.com/rhpds/showroom-intelligent-apps-rag