Back to changelog

Cerebras: Ultra-Fast Inference with 6 New Models

New Cerebras provider with six high-performance models, including GPT-OSS 120B and Qwen 3, now available through LLM Gateway.

Cerebras: Ultra-Fast Inference with 6 New Models

We're excited to announce support for Cerebras as a new provider in LLM Gateway, offering ultra-fast, high-throughput inference with six powerful models.

Cerebras is available via the LLM Gateway with the provider ID cerebras. Learn more about the Cerebras inference platform at cerebras.ai.

🎯 New Cerebras Models

Cerebras models

πŸš€ Getting Started with Cerebras

All Cerebras models are available via the OpenAI-compatible chat completions API:

1curl -X POST https://api.llmgateway.io/v1/chat/completions \2  -H "Authorization: Bearer $LLM_GATEWAY_API_KEY" \3  -H "Content-Type: application/json" \4  -d '{5    "model": "cerebras/gpt-oss-120b",6    "messages": [{"role": "user", "content": "Explain how Cerebras inference works"}]7  }'

Try Cerebras models in the Playground πŸš€

Get started now πŸš€