Skip to main content

Supported Providers

BoostGPT supports 9 AI providers with 50+ models spanning from ultra-fast nano models to advanced reasoning models.

Key Features

Use your own API keys for OpenAI, Google, Anthropic, xAI, DeepSeek, Mistral, Cohere, Groq, or Ollama. You control costs and rate limits.
Don’t have API keys? Use our pooled infrastructure on paid plans. We handle provisioning, rate limits, and scaling.
Access advanced reasoning models like O1, O3 Mini, DeepSeek R1, Gemini Flash Thinking, and Claude Extended Thinking for complex problem-solving.
Choose models based on speed, cost, reasoning capability, and context window. Use fast models for simple queries, reasoning models for complex tasks.
Run models locally using Ollama. Keep data private, eliminate API costs, and use custom fine-tuned models.

Model Categories

Speed-Optimized (Nano/Mini)

Ultra-fast models for simple tasks, high-volume applications, and real-time responses. Best for: FAQs, basic chat, high-traffic bots
Examples: GPT-4.1 Nano, Llama 3.1 Instant (8B), Gemini 2.5 Flash Lite

Balanced (Standard)

Great all-around models balancing speed, cost, and capability. Best for: Most production use cases, customer support, content generation
Examples: GPT-4o Mini, Claude Sonnet 4.5, Gemini 2.5 Flash

Advanced (Pro/Large)

Powerful models for complex tasks requiring deep understanding. Best for: Complex queries, creative writing, technical analysis
Examples: GPT-5, Claude Opus 4.1, Gemini 2.5 Pro

Reasoning Models

Specialized models with extended thinking for complex problem-solving. Best for: Math, coding, logic puzzles, multi-step reasoning
Examples: O1, DeepSeek R1, Gemini Flash Thinking, Claude Extended Thinking
Learn more about reasoning models →

Model Comparison

ProviderModelSpeedCostReasoningContextCredits
OpenAIO1SlowHighExceptional200K6
OpenAIGPT-5MediumHighExceptional200K5
OpenAIGPT-4oMediumMediumExcellent128K5
OpenAIGPT-4o MiniFastLowGood128K1
GoogleGemini 2.5 ProSlowHighExceptional2M3
GoogleGemini 2.5 FlashFastMediumExcellent1M2
AnthropicClaude Opus 4.1MediumHighExceptional200K5
AnthropicClaude Sonnet 4.5MediumMediumExcellent200K3
xAIGrok 3 ReasoningMediumMediumExcellent128K4
DeepSeekDeepSeek R1MediumMediumExcellent64K2
GroqLlama 3.3 (70B)FastMediumExcellent128K3
OllamaLlama 3.3 (70B)MediumFree*Excellent128K2
*Local hosting costs (compute) not included View full comparison →

Using Your Own Models (Ollama)

The only way to use custom models in BoostGPT is through Ollama.
1

Install Ollama

Download and install Ollama on your server or local machine.
2

Pull Your Model

    ollama pull llama3.3:70b
    # Or your custom fine-tuned model
3

Configure BoostGPT

Set your Ollama host URL in the dashboard or via SDK:
    const bot = await client.chat({
      bot_id: 'my-bot-id',
      provider_host: 'http://localhost:11434',  // or your server IP
      model: 'llama3.3:70b',
      message: 'Hello, world!' 
    });
4

Deploy

Your agent will now use your self-hosted Ollama model. No API costs, full control.
Learn more about Ollama →

Credits System

Each model has a credit cost per message. Credits are consumed when your agent generates responses.
  • Nano/Mini models: 0.5-1 credits (cheapest)
  • Standard models: 2-3 credits
  • Pro/Large models: 4-5 credits
  • Reasoning models: 3-6 credits (most expensive, but highest quality)
Credits are only for agent responses. Incoming messages, training data, and API calls don’t consume credits.

Choosing the Right Model

For Creators (No-Code Dashboard)

  1. Go to your bot settings
  2. Click “Model Selection”
  3. Choose based on your needs:
    • Fast responses → GPT-4o Mini, Gemini Flash Lite
    • Best quality → GPT-5, Claude Opus, Gemini Pro
    • Complex reasoning → O1, DeepSeek R1
    • Low cost → Use nano/mini variants

For Developers (SDK/API)

const bot = await client.createBot({
  name: 'Support Bot',
  model: 'gpt-5.1', // Fast & affordable
  // Or for complex queries:
  // model: 'o1', // Advanced reasoning
});

Provider-Specific Setup

Next Steps