Skip to main content

Overview

Groq provides blazing-fast inference for popular open-source models through their custom LPU hardware. Known for exceptional speed while maintaining strong model quality.

Available Models

DeepSeek Llama (70B)

2 credits • Distilled reasoning model
  • 64K context window
  • Excellent logical capabilities
  • Speed: Fast • Cost: Low
  • Reasoning model optimized for efficiency

Llama 3.3 Versatile (70B)

3 credits • Versatile large model
  • 128K context window
  • Excellent reasoning across diverse tasks
  • Speed: Fast • Cost: Medium
  • Best for: General-purpose applications

Llama 3.1 Instant (8B)

1 credit • Ultra-fast small model
  • 128K context window
  • Good reasoning with minimal latency
  • Speed: Extremely Fast • Cost: Very Low
  • Best for: Real-time applications

Qwen QWQ (32B)

2 credits • Multilingual reasoning
  • 32K context window
  • Excellent reasoning and multilingual
  • Speed: Fast • Cost: Low
  • Reasoning model for diverse tasks

Qwen 3 (32B)

2 credits • Next-gen instruction model
  • 32K context window
  • Excellent instruction-following
  • Speed: Fast • Cost: Low
  • Best for: Strong language tasks

Mistral Saba (24B)

2 credits • Reasoning and chat
  • 32K context window
  • Excellent reasoning
  • Speed: Fast • Cost: Low
  • Best for: Balanced performance

LLaMA 4 Maverick (17B 128E)

3 credits • 128 experts model
  • 128K context window
  • Excellent for advanced instructions
  • Speed: Medium • Cost: Medium
  • Best for: Complex instruction tasks

LLaMA 4 Scout (17B 16E)

1 credit • 16 experts model
  • 128K context window
  • Good for efficient inference
  • Speed: Fast • Cost: Low
  • Best for: Lighter workloads

Setup

Using BoostGPT-Hosted API Keys

1

Select Groq Model

In your BoostGPT dashboard, select any Groq model when creating or configuring your bot.
2

Choose Your Model

  • Llama 3.1 Instant: For real-time, ultra-fast responses
  • Llama 3.3 Versatile: For balanced performance
  • DeepSeek Llama 70B: For reasoning tasks
  • Qwen models: For multilingual applications

Using Your Own Groq API Key

1

Navigate to Integrations

Go to app.boostgpt.co and select Integrations
2

Select Groq

Find and click on the Groq provider
3

Add API Key

Enter your Groq API key and select which agents will use this key
4

Save Configuration

Click save to apply your custom API key

Performance Benefits

Groq’s LPU (Language Processing Unit) hardware provides inference speeds up to 10x faster than traditional GPUs, making it ideal for real-time applications.

Next Steps