Groq Mixtral 8x7B

Groq

Mixtral MoE on Groq's LPU. Fast, cost-effective inference for general tasks.

Context Window: 32K tokens Released: 2024-03 Best For: High-throughput text generation, chat, fast responses
  • MoE architecture
  • Ultra-fast
  • Low cost
  • Groq Mixtral 8x7B Pricing

    Token TypePrice per Million
    Input tokens$0.240
    Output tokens$0.240

    Estimated Cost by Project Size

    Realistic cost estimates for common coding scenarios. Assumes 30% cache hit rate where caching is available.

    ScenarioToken UsageEstimated Cost
    Small Script (1K lines) 50K input / 30K output $0.02
    Medium Feature (10K lines) 500K input / 200K output $0.13
    Large Project (50K lines) 2,500K input / 1,000K output $0.66
    Code Review (5K lines) 250K input / 25K output $0.05

    Get Access to Groq Mixtral 8x7B

    Ready to start using Groq Mixtral 8x7B? Get API access directly from Groq.

    Get API Access → Try Groq Mixtral 8x7B Free →

    How Does Groq Mixtral 8x7B Compare?

    ModelInput ($/M)Medium Feature Cost
    Groq Mixtral 8x7B $0.240 $0.13 selected
    Claude 3 Haiku $0.250 $0.34 Compare
    Qwen Coder Turbo $0.250 $0.34 Compare
    Llama 3.3 70B $0.250 $0.29 Compare
    DeepSeek Chat V3 $0.270 $0.31 Compare
    DeepSeek Coder V2 $0.270 $0.31 Compare

    Related Models

    Categories