Skip to main content

Welcome to LlamaFarm

LlamaFarm brings enterprise AI capabilities to everyone. Run powerful language models, document processing, and intelligent retrieval—all locally on your hardware. No cloud required. No data leaves your machine.

Found a bug or have a feature request?

Why LlamaFarm?

🔒 Edge AI for Everyone

Run sophisticated AI workloads on your own hardware:

  • Complete Privacy — Your documents, queries, and data never leave your device
  • No API Costs — Use open-source models without per-token fees
  • Offline Capable — Works without internet once models are downloaded
  • Hardware Optimized — Automatic GPU/NPU acceleration on Apple Silicon, NVIDIA, and AMD

🧠 Production-Ready AI Stack

LlamaFarm isn't just a wrapper—it's a complete AI development platform:

CapabilityWhat It Does
RAG (Retrieval-Augmented Generation)Ingest PDFs, docs, CSVs and query them with AI. Your documents become searchable knowledge.
Multi-Model RuntimeSwitch between Ollama, OpenAI, vLLM, or local GGUF models in one config file.
Custom ClassifiersTrain text classifiers with 8-16 examples using SetFit. No ML expertise required.
Anomaly DetectionDetect outliers in logs, metrics, or transactions with one API call.
OCR & Document ExtractionExtract text and structured data from images and PDFs.
Named Entity RecognitionFind people, organizations, and locations in your text.
Agentic Tools (MCP)Give AI models access to filesystems, databases, and APIs.

⚡ Developer Experience

  • Config-Driven — Define your entire AI stack in llamafarm.yaml
  • CLI + Web UI — Use the lf command line or the Designer visual interface
  • REST API — OpenAI-compatible endpoints for easy integration
  • Extensible — Add custom parsers, embedders, and model providers

Get Started in 60 Seconds

Option 1: Desktop App (Easiest)

Download the all-in-one desktop application:

The desktop app bundles everything: server, Universal Runtime, and the Designer web UI.

Option 2: CLI Installation

Install the lf command-line tool:

macOS / Linux:

curl -fsSL https://raw.githubusercontent.com/llama-farm/llamafarm/main/install.sh | bash

Windows (PowerShell):

irm https://raw.githubusercontent.com/llama-farm/llamafarm/main/install.ps1 | iex

Or download directly:

Verify installation:

lf --help

📺 See It In Action

Quick Overview (90 seconds): https://youtu.be/W7MHGyN0MdQ

Complete Walkthrough (7 minutes): https://youtu.be/HNnZ4iaOSJ4


What Can You Build?

Document Q&A

Upload your company's documents and ask questions in natural language:

lf datasets upload knowledge-base ./contracts/*.pdf
lf datasets process knowledge-base
lf chat "What are our standard payment terms?"

Custom Intent Classification

Train a classifier to route support tickets:

# Train with just 8 examples per category
POST /v1/ml/classifier/fit
{
"model": "ticket-router",
"training_data": [
{"text": "I can't log in", "label": "auth"},
{"text": "Charge me twice", "label": "billing"},
...
]
}

Real-Time Anomaly Detection

Monitor API logs for suspicious activity:

# Train on normal traffic
POST /v1/ml/anomaly/fit
{"model": "api-monitor", "data": [...normal_requests...]}

# Detect anomalies in real-time
POST /v1/ml/anomaly/detect
{"model": "api-monitor", "data": [...new_requests...]}

Document Processing Pipeline

Extract structured data from invoices and forms:

curl -X POST http://localhost:8000/v1/vision/ocr \
-F "file=@invoice.pdf" \
-F "model=surya"

Choose Your Path

Get StartedGo DeeperBuild Your Own
Quickstart — Install, init, chat, ingest your first datasetCore Concepts — Architecture, sessions, and componentsExtending LlamaFarm — Add runtimes, stores, parsers
Designer Web UI — Visual interface for project managementConfiguration Guide — Schema-driven project settingsRAG Guide — Strategies, processing pipelines
CLI Reference — Command matrix and examplesModels & Runtime — Configure AI models and providersAPI Reference — Full REST API documentation

Philosophy

  • Local-first, cloud-aware — Everything works offline, yet you can point at remote runtimes when needed
  • Configuration over code — Projects are reproducible because behavior lives in llamafarm.yaml
  • Composable modules — RAG, prompts, and runtime selection work independently but integrate cleanly
  • Edge for everyone — Enterprise AI capabilities without enterprise infrastructure
  • Open for extension — Add custom providers, stores, and utilities

Advanced: MCP (Model Context Protocol)

LlamaFarm supports MCP for giving AI models access to external tools like filesystems, databases, and APIs.

mcp:
servers:
- name: filesystem
transport: stdio
command: npx
args: ["-y", "@modelcontextprotocol/server-filesystem", "/path/to/dir"]

runtime:
models:
- name: assistant
provider: openai
model: gpt-4
mcp_servers: [filesystem]

Learn more about MCP →


Ready to build? Start with the Quickstart.