Welcome to LlamaFarm
LlamaFarm helps you ship retrieval-augmented and agentic AI apps from your laptop to production. It is fully open-source and intentionally extendable—swap model providers, vector stores, parsers, and CLI workflows without rewriting your project.
🖥️ Desktop App
Get started instantly with our desktop application — no command line required.
The desktop app includes everything you need: visual project management, dataset uploads, chat interface, and built-in model management. See hardware requirements →
📺 Video Demo
Quick Overview (90 seconds): https://youtu.be/W7MHGyN0MdQ
Get a fast introduction to LlamaFarm's core features and see it in action.
What You Can Do Today
- Prototype locally with Ollama or any OpenAI-compatible runtime (vLLM, Together, custom gateways).
- Ingest and query documents using configurable RAG pipelines defined entirely in YAML.
- Choose your interface – Use the powerful
lfCLI for automation and scripting, or the Designer web UI for visual project management with drag-and-drop dataset uploads and interactive configuration. - Extend everything from model handlers to data processors by updating schemas and wiring your own implementations.
- Give models superpowers with MCP (Model Context Protocol) – Connect your AI to local tools, APIs, and databases through a standardized protocol.
MCP (Model Context Protocol)
LlamaFarm supports MCP – a standardized protocol for giving AI models access to external tools. Connect your AI to filesystems, databases, APIs, and custom business logic.
mcp:
servers:
- name: filesystem
transport: stdio
command: npx
args: ['-y', '@modelcontextprotocol/server-filesystem', '/path/to/dir']
runtime:
models:
- name: assistant
provider: openai
model: gpt-4
mcp_servers: [filesystem]
Key features:
- Per-model access control – Different models get different tools
- Multiple transports – STDIO (local), HTTP (remote), SSE (streaming)
- Persistent sessions – Efficient connection management
Choose Your Own Adventure
| Get Started | Go Deeper | Build Your Own |
|---|---|---|
| Quickstart – install, init, chat, ingest your first dataset. | Core Concepts – architecture, sessions, and components. | Extending LlamaFarm – add runtimes, stores, parsers, and CLI commands. |
| Designer Web UI – visual interface for project management. | Configuration Guide – schema-driven project settings. | RAG Guide – strategies, processing pipelines, and monitoring. |
| CLI Reference – command matrix and examples. | Models & Runtime – configure AI models and providers. | Prompts – prompt engineering and management. |
Philosophy
- Local-first, cloud-aware – everything works offline, yet you can point at remote runtimes when needed.
- Configuration over code – projects are reproducible because behaviour lives in
llamafarm.yaml. - Composable modules – RAG, prompts, and runtime selection work independently but integrate cleanly.
- Flexible interfaces – Use the CLI for automation, the Designer for visual management, or the REST API for custom integrations.
- Open for extension – documentation includes patterns for registering new providers, stores, and utilities.
The Designer Web UI provides a browser-based interface for managing projects, uploading datasets, and testing your AI—all without touching the command line. It's automatically available at http://localhost:8000 when you run lf start. Learn more →
🎥 In-Depth Tutorial
Complete Walkthrough (7 minutes): https://youtu.be/HNnZ4iaOSJ4
Watch a comprehensive demonstration of LlamaFarm's features including project setup, dataset ingestion, RAG queries, and configuration options.
Ready to build? Start with the Quickstart and keep the CLI open in another terminal.