Introduction

Welcome to this hands-on workshop on building agentic AI applications using Llama Stack, LangGraph, Langfuse, Langflow, FastAPI and FastMCP on Red Hat OpenShift AI. This lab series takes you from deploying your first Llama Stack instance to building AI agents that interact with enterprise systems through the Model Context Protocol (MCP).

lab arch flow

What you'll learn

Deploy and explore Llama Stack

You'll deploy a Llama Stack distribution on OpenShift, discover its comprehensive API landscape, and interact with language models through multiple inference interfaces including the native Response API.

Implement RAG capabilities

Build Retrieval-Augmented Generation (RAG) systems using Llama Stack's built-in vector stores, embedding models, and file search tools to ground AI responses in your own documents.

Apply safety guardrails

Learn how to implement content moderation and safety shields to ensure your AI applications meet enterprise security and compliance requirements.

Web Search Tool

Leverage Tavily to augment context and address your model’s knowledge cutoff and limited training data.

Integrate business systems via MCP

Deploy and register Model Context Protocol servers that bridge Llama Stack and LangGraph agents with backend microservices, enabling AI to access customer data, financial transactions, and other enterprise capabilities.

Build autonomous agents

Create intelligent agents using both the native Llama Stack Client and popular frameworks like LangGraph that can reason about tool usage, execute multi-step workflows, and provide natural language interfaces to complex business functions.

Deploy production applications

Complete the journey by deploying a full-stack AI agent application with a web-based chat interface that brings together inference, tools, MCP, traces, evals and feedback.

Prerequisites

This workshop assumes basic familiarity with:

  • Linux command line (terminal usage) and bash scripting

  • Python programming fundamentals

  • REST APIs and JSON

  • Kubernetes/OpenShift concepts (pods, services, routes, configmaps)

Command line tools included in your environment:

This workshop comes with a provided terminal (Showroom Terminal) that contains already all the tools you need such as:

  • python

  • pip

  • git

  • curl, source, echo, export, sed, awk, grep

  • jq

  • oc

  • helm

  • openssl

  • watch

You have two terminals on the right side of this page, Terminal 1 and Terminal 2. While you can use just Terminal 1 for most of the lab, sometimes it’s comfortable to use both while deploying is in progress or when you have two processes to start.

Lab architecture

lab diagram

Throughout this workshop, you'll work with an enterprise AI architecture running on OpenShift:

  • Llama Stack Distribution: The core AI platform providing unified APIs for inference, agents, RAG, and tools

  • vLLM: High-performance model serving using the Qwen3 14B model, Llama Scout 17B, Llama Guard 3 1B, etc.

  • MCP Servers: Protocol translation layer connecting agents to backend systems

  • Backend Microservices: PostgreSQL-backed REST APIs providing customer and finance capabilities

  • Agent Applications: Python applications demonstrating various agent architectures and frameworks

All components run on Red Hat OpenShift AI 3, leveraging enterprise-grade infrastructure for production AI workloads.

Workshop Modules

  1. Introduction (this module) - Workshop overview and objectives

  2. Deploying Llama Stack - Install and configure your AI platform

  3. Exploring Llama Stack - Discover APIs and test inference capabilities

  4. RAG - Implement document retrieval and knowledge grounding Optional

  5. Evals - Evaluate the responses from models Optional

  6. Shields - Apply safety and content moderation Optional

  7. Web Search - Using Tavily to augment context with web results Optional

  8. Backend Setup - Deploy microservices and MCP servers

  9. MCP - Build agents with Python that use business function tools via MCP

  10. Agent - Explore the Llama Stack Agent API

  11. Agent - Deploy a complete LangGraph-based agent application

  12. Traces - Using Langfuse for traces, evals and feedback

  13. Workbench - RHOAI has an in-cluster, in-browser VS Code Optional

By the end of this workshop, you'll have hands-on experience building, deploying, and operating agentic AI systems on OpenShift using Llama Stack, FastMCP, LangGraph, FastAPI, Langfuse, and enterprise infrastructure.

Let's begin!