🦙 Local Model Support
OrcBot v2.1 introduces first-class support for local LLMs, allowing you to run your agent with total privacy and zero per-token costs.
Manage AI Models -> Ollama / Local Models.
Supported Providers
OrcBot's local model engine is designed around the OpenAI-compatible API standard. While Ollama is the primary supported platform, any server that exposes a /v1/chat/completions endpoint will work.
- Ollama (Native integration with auto-start and model pull)
- LM Studio
- vLLM
- LocalAI
- Text Generation WebUI
Configuration
To use a local model, update your configuration using one of the methods below.
Using the TUI (Recommended)
- Run
orcbot ui - Navigate to Manage AI Models
- Select Ollama / Local Models
- Follow the prompts to start the server, pull a model, and set it as primary.
Using orcbot.config.yaml
llmProvider: ollama
ollamaApiUrl: http://localhost:11434
modelName: llama3
Using Environment Variables
ORCBOT_LLM_PROVIDER=ollama
ORCBOT_OLLAMA_API_URL=http://localhost:11434
ORCBOT_MODEL_NAME=llama3
Native Tool Calling
One of OrcBot's core strengths is its ability to use tools (shell, browser, etc.). Unlike many local agent implementations that rely on brittle text parsing, OrcBot uses Native Tool Calling with local models that support it.
When a tool-capable local model (like llama3 or mistral) is used, OrcBot sends the tool definitions as part of the structured API request, ensuring high reliability and complex task execution.
Performance Tuning
Running models locally requires significant hardware resources. Here are some tips for optimizing performance:
GPU Acceleration
Always ensure your GPU drivers are up to date. Ollama will automatically use NVIDIA (CUDA) or Apple (Metal) acceleration if available.
Model Sizing
For complex strategic planning, use at least a 7B parameter model. For better tool usage and reasoning, 13B or 70B models are recommended if your VRAM allows.
Security & Air-Gapping
By combining Local Model support with OrcBot's User-Space Restricted Shell, you can create a highly secure, air-gapped autonomous system. Since the model runs on your machine, sensitive data like logs, file paths, and internal reasoning never leave your local network.