Changelog
Stay up to date with the latest features, improvements, and fixes in LlamaFarm.
Latest Release
v0.0.27 — 2026-02-16
LlamaFarm 0.0.27: Addons, Smarter RAG, and Runtime Resilience
This release introduces the addons system, smarter RAG defaults, and significant runtime stability improvements.
New Features
Addons System
LlamaFarm now supports addons — modular extensions you can install and enable to expand your platform's capabilities. The Designer includes a polished UX for browsing, installing, and managing addons, with sequential installation and auto-enable on install for a smooth experience.
Per-Model RAG Defaults
You can now configure default RAG retrieval strategies on a per-model basis. This means different models can automatically use the retrieval settings that work best for them — no manual configuration needed each time.
RAG Source Chunks in Test Outputs
The Designer now shows RAG source chunks directly in test outputs, so you can see exactly which documents your model is referencing. Great for debugging retrieval quality and understanding model responses.
Cascading Data Processing Strategies
The server now supports cascading default data processing strategies, making it easier to set up sensible defaults that flow through your entire pipeline.
Anomaly Detection Documentation
Comprehensive docs, use-cases, and a full demo for the anomaly detection feature introduced in v0.0.24 — making it much easier to get started with outlier detection.
Infrastructure
- Binary component builds for faster CI and distribution
- Server port change — default port moved from 8000 to 14345 to avoid conflicts
Bug Fixes
- Smart GPU allocation — prevents multi-model OOM crashes by intelligently managing GPU memory across loaded models
- Event loop protection — model loading in the Universal Runtime no longer blocks the event loop, improving responsiveness during heavy loads
- API system prompts — fixed a bug where API-provided system prompts were being overridden by config-level system prompts
- Designer improvements — better delete UX, ghost project handling, fixed 404 on train button, improved onboarding checklist updates after demo project conversion
- Audio error handling — improved error handling in the Designer for audio features
v0.0.26 — 2026-01-27
LlamaFarm 0.0.26: Smarter, Faster, and More Accessible
This release brings a range of improvements to make LlamaFarm more intuitive, efficient, and accessible across different platforms and use cases.
New Features and Enhancements
Reusability and Configuration Improvements
We've introduced reusable components in the configuration system, allowing you to define and reuse common settings across different parts of your application. This makes managing complex configurations much simpler and reduces duplication.
Enhanced RAG Capabilities
Universal RAG - We've added zero-config default strategies that work out of the box for most use cases. No more complex setup required to get started with retrieval-augmented generation.
Document Preview - You can now preview documents with strategy selection directly in the Designer, making it easier to understand how your RAG pipeline processes different file types.
Dataset Management
New sample datasets for gardening and home repair scenarios help you get started quickly with realistic data. Plus, datasets now auto-process on upload, eliminating manual processing steps.
Developer Experience
Dynamic Value Substitution - Prompts and tools now support dynamic variable substitution, making your configurations more flexible and powerful.
Service Status Panel - A new status panel in the Designer header gives you real-time visibility into your LlamaFarm services, so you know exactly what's running.
Audio and Speech
This release introduces a full-duplex speech reasoning pipeline with audio processing capabilities in the Universal Runtime. Build voice-enabled AI applications with ease.
Cross-Platform Support
- Desktop App Improvements - Better splash screen UX and enhanced cross-platform support
- Intel Mac Support - Added support for Intel Macs (x86_64) with PyTorch 2.2.2
- Jetson/Tegra Optimization - Improved CUDA optimization and unified memory GPU support
Bug Fixes
- Fixed dev builds stopping running services
- Resolved sample project creation failures
- Fixed chat input clearing during streaming
- Improved error display and Service Status panel reliability
Recent Releases
v0.0.25 — 2026-01-14
LlamaFarm 0.0.25: Native Tool Calling and Developer Productivity
This release focuses on improving the developer experience with better tooling, native tool calling support, and automatic file processing capabilities.
New Features
Native Tool Calling
The Universal Runtime now supports native tool calling, enabling your AI models to interact with external tools and APIs more efficiently. This is a major step forward for building agentic AI applications that can take actions in the real world.
Automatic File Processing
Files uploaded to datasets now process automatically, eliminating the manual processing step and streamlining your workflow. Just upload and go.
Enhanced Designer Development Tools
The Designer now includes comprehensive API call logging in the dev tools panel, making it easier to debug and understand how your application communicates with the backend. See every request and response in real-time.
Streaming Model Downloads
Embedding model downloads now use SSE streaming, providing real-time progress updates so you always know exactly what's happening during long downloads.
Extended Testing Capabilities
The test space now includes support for anomaly detection and classifier tests, giving you more ways to validate your AI models before deployment.
Bug Fixes
- Fixed config validation error output for clearer debugging
- Resolved install and run failures on Windows with NVIDIA GPUs
- Removed parser fallback to prevent unexpected behavior
- Enabled offline GGUF model loading for air-gapped environments
v0.0.24 — 2026-01-06
LlamaFarm 0.0.24: Anomaly Detection
This release introduces anomaly detection capabilities to help identify outliers and unusual patterns in your data.
New Features
Anomaly Detection
The Universal Runtime now supports anomaly detection with configurable normalization methods for scoring. Whether you're monitoring for fraud, equipment failures, or data quality issues, LlamaFarm can now help identify when something doesn't look right.
Designer UX for Anomaly Detection
The Designer includes a new interface for configuring and testing anomaly detection models, making it easy to set up detection pipelines and visualize results.
Bug Fixes
- Fixed anomaly and classifier UX issues in the Designer for smoother workflows
v0.0.23 — 2025-12-20
LlamaFarm 0.0.23: Stability Improvements
A focused stability release addressing a critical logging issue in the Universal Runtime.
Bug Fixes
- Fixed broken pipe errors caused by problematic logging in the Universal Runtime, improving reliability for long-running inference tasks
v0.0.22 — 2025-12-19
LlamaFarm 0.0.22: Inference Fix
A quick bug fix release addressing an issue with logits processor handling.
Bug Fixes
- Fixed logits_processor to be passed as callable instead of list, resolving inference issues with certain model configurations
v0.0.21 — 2025-12-19
LlamaFarm 0.0.21: Specialized ML Models and Vision API
A feature-packed holiday release bringing specialized ML models, vision capabilities, and major Designer enhancements.
New Features
Specialized ML Models
Added support for OCR, document extraction, and anomaly detection models in the Universal Runtime. These specialized models expand what you can build with LlamaFarm beyond text generation - now you can extract text from images, parse documents, and detect anomalies.
Vision API
New vision router and model versioning for ML endpoints, enabling image understanding capabilities in your applications. Build apps that can see and understand visual content.
Designer Improvements
- Santa's Holiday Helper Demo - A festive demo project to help new users get started
- Enhanced RAG UX - Improved retrieval strategy settings in test chat
- Data Enhancements - Better tools for managing your datasets
- Global Project Listing - Easily see all your projects in one place
Cross-Platform Support
Native llama-cpp bindings now included for all platforms, and Windows builds correctly include the .exe extension for seamless installation.
Bug Fixes
- Fixed upgrade failures on Linux
- Ensured multi-arch Linux builds work correctly
- Fixed model unload cleanup and OpenAI message validation
- Removed console log spam in Designer
Older Releases
View all releases
v0.0.20 — 2025-12-10
Auto-Start Services, RAG Stats, and Reliability Improvements
New Features
- Auto-Start Service Flag - Services can now start automatically when you run LlamaFarm
- More GGUF Download Options - More quantization options for model downloads in Designer
- RAG Database Listing - List all documents in your RAG databases
- RAG Statistics - View detailed stats about your RAG setup
- Chunk Cleanup - Automatically remove database chunks when files are deleted
- Data Processing Control - Start and stop data processing from the API
Bug Fixes
- Fixed first-run startup failures for new users
- Improved path resolution with
~expansion - Better process manager locking to prevent conflicts
- Fixed upgrade hang caused by process stop deadlock
- Prevented storage of failed vectors in RAG
v0.0.19 — 2025-12-03
Automatic Model Downloads, Custom RAG Queries, and Reasoning Models
New Features
- Automatic Model Download Management - Models download automatically when needed
- Custom RAG Queries - Send custom RAG queries through the chat/completions endpoint
- Thinking/Reasoning Model Support - Support for models that show their reasoning process
- Database CRUD API - Full create, read, update, delete operations for databases
- Better Day-2 UX - Improved experience for returning users
- Disk Space Checking - Check available disk space before downloading models
- GGUF Model Listing - Browse available GGUF models for download
Bug Fixes
- Fixed datasets endpoint trailing slash requirement
- Improved cross-filesystem data moves
- Fixed PDF parsing issues in RAG
- Addressed demo timeout issues
v0.0.18 — 2025-11-25
v0.0.17 — 2025-11-24
Bug Fixes and Documentation
Bug Fixes
- Fixed empty prompts array for new projects
- Added troubleshooting documentation
- Fixed HuggingFace progress bar crashes
v0.0.16 — 2025-11-23
v0.0.15 — 2025-11-22
Desktop App Launch and GGUF Model Support
New Features
- Desktop App - Full Electron desktop app with auto-updates and polished UI
- GGUF Model Support - Run quantized GGUF models in the Universal Runtime
- Demo Project System - Interactive demo projects to help new users get started
- Universal Event Logging - Comprehensive observability across the platform
- Enhanced Tool Calling - Improved tool calling capabilities
- Project Cloning - Create new projects from existing ones
Bug Fixes
- Fixed upgrade failures on Unix-like systems
- Improved RAG integration and chat context management
- Fixed database tab switching in Designer
- Better dataset validation and status display
v0.0.14 — 2025-11-13
v0.0.13 — 2025-11-11
v0.0.12 — 2025-11-11
Project Management and Config Editor
New Features
- Delete Projects - Remove projects from CLI and API
- Config Editor Enhancements - Copy button, search, anchor points, unsaved changes prompts
- Embedding Strategies API - Configure embedding strategies via API
- MCP Server Config - Add MCP server configuration to runtime
- Project Context Provider - Better project context management
v0.0.11 and earlier
For releases v0.0.11 and earlier, please see the full changelog on GitHub.
About These Release Notes
These release notes are generated from our conventional commit history. For the complete structured changelog with commit links and PR references, see the CHANGELOG.md on GitHub.
Stay Updated
- GitHub Releases: github.com/llama-farm/llamafarm/releases
- Reddit: r/LlamaFarm
- Discord: Join our community