Closing

Congratulations on completing this hands-on workshop on building agentic AI applications using Llama Stack, LangGraph, Langfuse, FastAPI and FastMCP on Red Hat OpenShift AI. You’ve journeyed from deploying your first Llama Stack instance to building AI agents that interact with enterprise systems through the Model Context Protocol (MCP).

lab arch flow

What you’ve learned

Deployed and explored Llama Stack

You deployed a Llama Stack distribution on OpenShift, discovered its comprehensive API landscape, and interacted with language models through multiple inference interfaces including the native Response API.

Implemented RAG capabilities

You built Retrieval-Augmented Generation (RAG) systems using Llama Stack’s built-in vector stores, embedding models, and file search tools to ground AI responses in your own documents.

Applied safety guardrails

You learned how to implement content moderation and safety shields to ensure your AI applications meet enterprise security and compliance requirements.

Leveraged Web Search Tools

You used Tavily to augment context and address your model’s knowledge cutoff and limited training data.

Integrated business systems via MCP

You deployed and registered Model Context Protocol servers that bridge Llama Stack and LangGraph agents with backend microservices, enabling AI to access customer data, financial transactions, and other enterprise capabilities.

Built autonomous agents

You created intelligent agents using both the native Llama Stack Client and popular frameworks like LangGraph that can reason about tool usage, execute multi-step workflows, and provide natural language interfaces to complex business functions.

Deployed production applications

You completed the journey by deploying a full-stack AI agent application with a web-based chat interface that brings together inference, tools, MCP, traces, evals and feedback.

Throughout this workshop, you worked with an enterprise AI architecture running on OpenShift:

  • Llama Stack Distribution: The core AI platform providing unified APIs for inference, agents, RAG, and tools

  • vLLM: High-performance model serving using the Qwen3 14B model, Llama Scout 17B, Llama Guard 3 1B, etc.

  • MCP Servers: Protocol translation layer connecting agents to backend systems

  • Backend Microservices: PostgreSQL-backed REST APIs providing customer and finance capabilities

  • Agent Applications: Python applications demonstrating various agent architectures and frameworks

All components ran on Red Hat OpenShift AI 3, leveraging enterprise-grade infrastructure for production AI workloads.

Workshop modules completed

  1. Introduction - Workshop overview and objectives

  2. Deploying Llama Stack - Installed and configured your AI platform

  3. Exploring Llama Stack - Discovered APIs and tested inference capabilities

  4. RAG - Implemented document retrieval and knowledge grounding

  5. Evals - Evaluated the responses from models

  6. Shields - Applied safety and content moderation

  7. Web Search - Used Tavily to augment context with web results

  8. Backend Setup - Deployed microservices and MCP servers

  9. MCP - Built agents with Python that use business function tools via MCP

  10. Agent - Explored the Llama Stack Agent API

  11. Agent - Deployed a complete LangGraph-based agent application

  12. Traces - Used Langfuse for traces, evals and feedback

  13. Langflow - Built graphical agent workflows

  14. Workbench - Explored OpenShift AI’s in-cluster, in-browser VS Code

You now have hands-on experience building, deploying, and operating agentic AI systems on OpenShift using Llama Stack,FastMCP, LangGraph, FastAPI, Langfuse, Langflow and enterprise infrastructure.

Next steps

Taking this to your own environment: The patterns you practiced — Llama Stack distribution, MCP servers, agent frameworks, observability with Langfuse — are all portable. You can reproduce this architecture on any OpenShift cluster with GPU access by following the same Helm charts and CRDs used in this lab. The key components (Llama Stack, vLLM, FastMCP, LangGraph, Langfuse, Langflow) are all open source.

Ready to continue your AI journey? Explore more innovative AI use cases and quickstart guides:

The AI quickstarts provide simple, focused examples designed for fast and easy deployment on the Red Hat AI platform, including:

  • Enterprise RAG Chatbot - Centralize company knowledge with retrieval-augmented generation

  • AI-powered Virtual Agent - Build conversational AI to automate customer interactions

  • Privacy-focused AI Assistant - Deploy healthcare AI with PII detection and content moderation

  • Self-Service IT Agents - Automate IT processes like laptop refresh workflows

  • HR Assistant - Replace hours of policy document searches with AI-powered assistance

  • Compliance Audit Systems - Deploy financial audit systems with real-time insights

  • Observability Analysis - Summarize and analyze AI model performance and cluster metrics

  • AI Product Recommendations - Transform e-commerce with AI-driven discovery

Each quickstart has been created by Red Hat experts and is designed to help you adopt and scale AI in your organization.

Thank you!

Thank you for participating in this workshop. We hope you found it valuable and that you’re excited to bring these agentic AI capabilities to your own projects and organization.

For questions, feedback, or support, please reach out to your Red Hat team or visit the Red Hat Customer Portal.

Happy building!