Skip to main content

LLM Providers

GenieBuilder provides a unified interface that lets you switch between cloud and local models without changing your workflow.

LLM Providers Settings

Provider Types

┌─────────────────────────────────────────────────────────┐
│ Provider Types │
├─────────────────────────────────────────────────────────┤
│ │
│ ┌──────────────┐ ┌──────────────┐ ┌──────────────┐ │
│ │ Cloud │ │ Local │ │ CLI Runners │ │
│ │ │ │ │ │ │ │
│ │ - OpenAI │ │ - Ollama │ │ - Claude │ │
│ │ - Anthropic │ │ - LM Studio │ │ - Kimi │ │
│ │ - Gemini │ │ - llama.cpp │ │ - Codex │ │
│ │ │ │ │ │ - Copilot │ │
│ │ API-based │ │ Self-hosted │ │ - Custom │ │
│ │ Scalable │ │ Privacy │ │ Agentic │ │
│ └──────────────┘ └──────────────┘ └──────────────┘ │
│ │
└─────────────────────────────────────────────────────────┘

Cloud Providers

OpenAI

Access GPT models through OpenAI's API.

Models:

  • GPT-5.4 (flagship)
  • GPT-5.4-mini (balanced)
  • GPT-5.4-nano (fast)
  • GPT-5.x series (when available)

Configuration:

Provider ID: openai
API Key: sk-...
Base URL: https://api.openai.com/v1 (default)

Best for:

  • General-purpose tasks
  • Complex reasoning
  • Code generation

Anthropic

Access Claude models through Anthropic's API.

Models:

  • Claude 4.6 (Opus, Sonnet)
  • Claude 4.5 Sonnet
  • Claude 3.5 series

Configuration:

Provider ID: anthropic
API Key: sk-ant-...
Base URL: https://api.anthropic.com (default)

Best for:

  • Long context (200K tokens)
  • Careful reasoning
  • Document analysis

Google Gemini

Access Gemini models through Google's API.

Models:

  • Gemini 3.1 Pro
  • Gemini 2.0 Flash
  • Gemini 1.5 series

Configuration:

Provider ID: google
API Key: AIza...
Base URL: https://generativelanguage.googleapis.com (default)

Best for:

  • Multimodal tasks
  • Large context (1M tokens)
  • Integration with Google services

Local Providers

Ollama

Run models locally with Ollama.

Setup:

# Install Ollama
brew install ollama

# Pull a model
ollama pull llama3
ollama pull codellama
ollama pull glm-4.7-flash

Configuration:

Provider ID: ollama
Base URL: http://127.0.0.1:11434/v1 (default)
API Key: (not required by default)

Best for:

  • Privacy (data stays local)
  • Offline use
  • Cost control
  • Experimentation

Recommended models:

  • llama3 — General purpose
  • codellama — Code generation
  • glm-4.7-flash — Fast responses
  • mistral — Balanced performance

LM Studio

Run models through LM Studio's local server.

Setup:

  1. Download LM Studio
  2. Load a model
  3. Start local server (default port 1234)

Configuration:

Provider ID: lm-studio
Base URL: http://localhost:1234/v1

Best for:

  • GUI model management
  • Easy model switching
  • Quantized models

llama.cpp

Direct integration with llama.cpp.

Setup:

# Build or download llama.cpp server
./server -m model.gguf -c 4096

Configuration:

Provider ID: llama.cpp
Base URL: http://localhost:8080

Best for:

  • Maximum control
  • Custom builds
  • Research/experimentation

CLI Runners

CLI Runners integrate external command-line AI agents. See CLI Runners for detailed documentation.

Claude Code

Anthropic's agentic coding tool.

Installation: npm install -g @anthropic-ai/claude-code

Features:

  • Rich tool use
  • File editing
  • Shell command execution
  • MCP integration

Kimi CLI

Moonshot AI's efficient CLI.

Features:

  • Fast responses
  • --yolo auto-execution mode
  • Clean JSON output
  • MCP support

OpenAI Codex

OpenAI's code review CLI.

Installation: npm install -g @openai/codex

Features:

  • Code review focus
  • --full-auto mode
  • JSON output format
  • MCP integration

GitHub Copilot CLI

GitHub Copilot for command line.

Features:

  • IDE-integrated experience
  • GitHub context awareness
  • MCP support

Provider Comparison

AspectCloudLocalCLI
SetupEasy (API key)Moderate (install)Moderate (install)
CostPer-tokenHardware onlyPer-use or free
PrivacyData sent to APILocal onlyVaries
OfflineNoYesDepends
QualityHighestGoodHigh
SpeedNetworkLocal HWVaries
ToolsNativeNativeVia MCP

Configuration

Adding a Provider

  1. Open SettingsProviders
  2. Select provider type
  3. Enter API key or base URL
  4. Test connection
  5. Save configuration

Per-Tab Provider Selection

Each chat tab can use a different provider:

  1. Open new chat tab
  2. Click provider dropdown
  3. Select provider and model
  4. Start conversation

Default Provider

Set a default provider for new tabs:

Settings → Chat → Default Provider

API Key Security

GenieBuilder uses the system's secure storage for API keys:

Supported Platforms

OSEncryption
macOSKeychain
WindowsDPAPI
LinuxSecret Service API / libsecret

Fallback

On unsupported systems, keys are stored in plain text with a visible warning.

Best Practices

  1. Use environment variables for CI/automation
  2. Rotate keys regularly
  3. Use restricted keys when possible (read-only, IP-limited)
  4. Never commit keys to version control

Model Selection Guide

Choose models based on your needs:

TaskRecommended
Code generationGPT-5.4, Claude 4.6, Codex CLI
Code reviewClaude 4.6, Codex CLI, Kimi
DocumentationClaude 4.6 (long context), GPT-5.4
Quick answersGPT-5.4-mini, local small models
Privacy-criticalLocal models (Ollama)
Offline workLocal models
Complex reasoningGPT-5.4, Claude 4.6 Opus
Agentic workflowsClaude Code, Kimi CLI

Troubleshooting

Connection Failed

If provider test fails:

  1. Verify API key is correct
  2. Check base URL for local providers
  3. Ensure network connectivity (cloud)
  4. Verify local server is running (local)

Slow Responses

If models respond slowly:

Cloud:

  • Check network connection
  • Try a different model (smaller = faster)
  • Check provider status page

Local:

  • Verify GPU acceleration is enabled
  • Reduce context window
  • Use a smaller quantized model

Context Length Errors

If you hit token limits:

  • Switch to model with larger context
  • Truncate conversation history
  • Split large inputs into chunks
  • Use Claude (200K) or Gemini (1M) for long documents

CLI Runner Not Found

If CLI runner fails:

  1. Verify CLI is installed and in PATH
  2. Check executable path in settings
  3. Test CLI directly in terminal
  4. Review CLI-specific requirements