Overview
OrcBot supports multiple LLM providers with automatic fallback and provider-specific model persistence. You can configure providers via environment variables or YAML configuration files.Quick Start
Provider Selection
Explicit LLM provider selection. Supported values:
openai- OpenAI (GPT-4, GPT-3.5, o1)google- Google Geminianthropic- Anthropic Claudebedrock- AWS Bedrockopenrouter- OpenRouter (200+ models)nvidia- NVIDIA NIMollama- Local Ollama models
The provider can also be inferred from the model name prefix (e.g.,
claude- → anthropic).Primary model to use. The provider is auto-detected from the model name if not explicitly set.
Per-provider model preferences. When you switch providers, OrcBot automatically restores the last model used for that provider.
Fallback models when primary provider fails:
Fast model for lightweight operations like summarization and classification.Examples:
gpt-4o-minigemini-1.5-flashclaude-3-5-haiku-20241022
OpenAI
Configuration
OpenAI API key (starts with
sk-).Policy: APPROVAL - API keys are sensitive and require approval for changes.Environment variable: OPENAI_API_KEYSupported Models
GPT-4o (Recommended)
GPT-4o (Recommended)
- Model ID:
gpt-4o - Context: 128K tokens
- Strengths: Best balance of intelligence, speed, and cost
- Use cases: General tasks, coding, reasoning
GPT-4o Mini
GPT-4o Mini
- Model ID:
gpt-4o-mini - Context: 128K tokens
- Strengths: Fast and affordable
- Use cases: Quick responses, summarization, classification
GPT-4 Turbo
GPT-4 Turbo
- Model ID:
gpt-4-turbo - Context: 128K tokens
- Strengths: Previous generation flagship
- Use cases: Complex reasoning, analysis
o1 Series (Reasoning)
o1 Series (Reasoning)
- Model IDs:
o1,o1-mini,o1-preview - Context: 128K tokens (o1), 128K tokens (o1-mini)
- Strengths: Advanced reasoning, complex problem-solving
- Limitations: No function calling, streaming only
- Use cases: Deep analysis, strategic planning
Example Configuration
orcbot.config.yaml
Google Gemini
Configuration
Google AI API key.Policy: APPROVAL - API keys are sensitive and require approval for changes.Environment variable:
GOOGLE_API_KEYGet your API key at: https://makersuite.google.com/app/apikeySupported Models
Gemini 2.0 Flash (Recommended)
Gemini 2.0 Flash (Recommended)
- Model ID:
gemini-2.0-flash-exp - Context: 1M tokens
- Strengths: Multimodal, fast, large context
- Use cases: General tasks, image analysis, long documents
Gemini 1.5 Pro
Gemini 1.5 Pro
- Model ID:
gemini-1.5-pro - Context: 2M tokens
- Strengths: Massive context window, high intelligence
- Use cases: Extremely long documents, complex analysis
Gemini 1.5 Flash
Gemini 1.5 Flash
- Model ID:
gemini-1.5-flash - Context: 1M tokens
- Strengths: Fast, affordable, good quality
- Use cases: Quick responses, summarization
Computer Use (Experimental)
Enable Gemini’s computer use capabilities for screen control.
Model to use for computer use features.
Example Configuration
orcbot.config.yaml
Anthropic Claude
Configuration
Anthropic API key (starts with
sk-ant-).Policy: APPROVAL - API keys are sensitive and require approval for changes.Environment variable: ANTHROPIC_API_KEYSupported Models
Claude 3.5 Sonnet (Recommended)
Claude 3.5 Sonnet (Recommended)
- Model ID:
claude-3-5-sonnet-20241022 - Context: 200K tokens
- Strengths: Best reasoning, coding, analysis
- Use cases: Complex tasks, code generation, writing
Claude 3.5 Haiku
Claude 3.5 Haiku
- Model ID:
claude-3-5-haiku-20241022 - Context: 200K tokens
- Strengths: Fast, affordable, intelligent
- Use cases: Quick responses, summarization
Claude 3 Opus
Claude 3 Opus
- Model ID:
claude-3-opus-20240229 - Context: 200K tokens
- Strengths: Highest intelligence, complex reasoning
- Use cases: Most demanding tasks, strategic planning
Example Configuration
orcbot.config.yaml
AWS Bedrock
Configuration
AWS region for Bedrock service (e.g.,
us-east-1, us-west-2).Environment variable: BEDROCK_REGION or AWS_REGIONAWS access key ID.Policy: LOCKED - Critical authentication credential.Environment variable:
BEDROCK_ACCESS_KEY_ID or AWS_ACCESS_KEY_IDAWS secret access key.Policy: LOCKED - Critical authentication credential.Environment variable:
BEDROCK_SECRET_ACCESS_KEY or AWS_SECRET_ACCESS_KEYAWS session token (for temporary credentials).Environment variable:
BEDROCK_SESSION_TOKEN or AWS_SESSION_TOKENSupported Models
Bedrock provides access to multiple model families:- Anthropic Claude:
anthropic.claude-3-sonnet-20240229-v1:0 - Amazon Titan:
amazon.titan-text-premier-v1:0 - Meta Llama:
meta.llama3-70b-instruct-v1:0 - Cohere Command:
cohere.command-r-plus-v1:0 - Mistral AI:
mistral.mistral-large-2407-v1:0
Example Configuration
orcbot.config.yaml
.env
OpenRouter
Configuration
OpenRouter API key.Policy: APPROVAL - API keys are sensitive and require approval for changes.Environment variable:
OPENROUTER_API_KEYGet your API key at: https://openrouter.ai/keysOpenRouter API base URL.Environment variable:
OPENROUTER_BASE_URLHTTP Referer header for OpenRouter requests (optional, for ranking).Environment variable:
OPENROUTER_REFERERApplication name for OpenRouter (optional, for ranking).Environment variable:
OPENROUTER_APP_NAMESupported Models
OpenRouter provides access to 200+ models from multiple providers:Popular Models
Popular Models
anthropic/claude-3.5-sonnet- Claude 3.5 Sonnetopenai/gpt-4o- GPT-4ogoogle/gemini-2.0-flash-exp:free- Gemini 2.0 Flash (free)meta-llama/llama-3.3-70b-instruct- Llama 3.3 70Bmistralai/mistral-large- Mistral Largedeepseek/deepseek-chat- DeepSeek V3
Free Models
Free Models
OpenRouter offers several free models:
google/gemini-2.0-flash-exp:freemeta-llama/llama-3.2-3b-instruct:freemicrosoft/phi-3-mini-128k-instruct:free
Free models have rate limits and may have lower availability.
Example Configuration
orcbot.config.yaml
Ollama (Local Models)
Configuration
Enable local Ollama integration for running models locally.
Ollama API endpoint URL.Policy: SAFE - Non-sensitive endpoint configuration.
Supported Models
Ollama supports running models locally:- Llama 3.3:
llama3.3:70b,llama3.3:8b - Qwen 2.5:
qwen2.5:72b,qwen2.5:32b - Mistral:
mistral:7b,mistral-nemo:12b - DeepSeek:
deepseek-r1:70b,deepseek-r1:8b - Phi-4:
phi4:14b - Gemma 2:
gemma2:27b,gemma2:9b
TUI Management
OrcBot’s TUI provides built-in Ollama management:- Auto-start server - Automatically starts Ollama if not running
- Pull models - Download models from the Ollama library
- Model selection - Switch between installed models
- OpenAI-compatible - Native tool calling support
Example Configuration
orcbot.config.yaml
Installation
Install Ollama
Download from https://ollama.com or install via package manager:
Additional Providers
OrcBot also supports these providers via direct API integration:NVIDIA NIM
NVIDIA NIM API key.Policy: APPROVAL - API keys are sensitive.Environment variable:
NVIDIA_API_KEYGroq
Groq API key for ultra-fast inference.Environment variable:
GROQ_API_KEYMistral AI
Mistral AI API key.Environment variable:
MISTRAL_API_KEYCerebras
Cerebras API key for fastest inference.Environment variable:
CEREBRAS_API_KEYxAI
xAI (Grok) API key.Environment variable:
XAI_API_KEYMulti-Provider Strategy
Configure multiple providers for automatic fallback:orcbot.config.yaml
When a provider fails, OrcBot automatically tries the fallback model, then switches to an alternative provider if configured.
Provider Auto-Detection
You can omitllmProvider and OrcBot will infer it from the model name:
Best Practices
Use Environment Variables
Store API keys in
.env files or environment variables, not in YAML configuration.Configure Fallbacks
Set fallback models to ensure continuity when primary provider fails.
Choose Fast Models
Configure
fastModelName for lightweight operations to reduce latency and cost.Test Locally First
Use Ollama for development and testing before using paid APIs.