WordSwarm AI: Reasoning & Benchmarking Workshop
Welcome to the WordSwarm AI workshop!
The live demo is a WordSwarm game with an AI agent that can play it autonomously.
What you’ll learn
In this workshop, you will:
-
Play with reasoning models — use
/thinkand/no-thinkprompting to control LLM reasoning behavior in a live AI agent -
Benchmark MaaS models — use GuideLLM to measure throughput, latency, and concurrency of models served via vLLM on OpenShift
-
Compare model performance — see how different models (kimi-k2-5, Nemotron, Llama, Gemma) perform on real-time tasks vs. synthetic benchmarks
Who this is for
This workshop is designed for AI/ML engineers, platform engineers, and developers who want hands-on experience with:
-
LLM reasoning capabilities (
/think,/no-think) -
Model benchmarking with GuideLLM
-
Model-as-a-Service (MaaS) on OpenShift with vLLM
Experience level: Beginner to Intermediate
