Skip to content

rh-aiservices-bu/llama-stack-tutorial

Repository files navigation

Llama Stack Tutorial

Welcome to the Llama Stack Tutorial - your comprehensive guide to learning and using Llama Stack.

What is Llama Stack?

Llama Stack is an open-source framework that makes it easier to build, run, and experiment with large language models on your own infrastructure. It provides:

  • Unified interface for different model providers (Ollama, vLLM, etc.)
  • Tools and agents for complex AI workflows
  • Safety and telemetry built-in
  • CLI, Python SDK, and web playground for development

Tutorial Content

This repository contains hands-on tutorials covering:

🟢 Beginner

  • Getting started with CLI and Python SDK
  • Using the web playground
  • Basic model interactions

🟡 Elementary

  • Model Context Protocol (MCP) integration
  • Retrieval Augmented Generation (RAG)
  • Building interactive agents

🟠 Intermediate

  • Safety and content filtering
  • Telemetry and observability
  • Production deployments

🔴 Advanced

  • Custom providers and extensions
  • Advanced agent workflows

Access Points

Contributing

This tutorial is designed to help developers learn Llama Stack through practical examples. The content is built with Antora and can be edited in the content/modules/ROOT/pages/ directory.


Releases

No releases published

Packages

No packages published

Contributors 4

  •  
  •  
  •  
  •