Cerebras: Ultra-Fast Inference with 6 New Models
New Cerebras provider with six high-performance models, including GPT-OSS 120B and Qwen 3, now available through LLM Gateway.

We're excited to announce support for Cerebras as a new provider in LLM Gateway, offering ultra-fast, high-throughput inference with six powerful models.
Cerebras is available via the LLM Gateway with the provider ID cerebras. Learn more about the Cerebras inference platform at cerebras.ai.
π― New Cerebras Models
π Getting Started with Cerebras
All Cerebras models are available via the OpenAI-compatible chat completions API:
1curl -X POST https://api.llmgateway.io/v1/chat/completions \2 -H "Authorization: Bearer $LLM_GATEWAY_API_KEY" \3 -H "Content-Type: application/json" \4 -d '{5 "model": "cerebras/gpt-oss-120b",6 "messages": [{"role": "user", "content": "Explain how Cerebras inference works"}]7 }'
1curl -X POST https://api.llmgateway.io/v1/chat/completions \2 -H "Authorization: Bearer $LLM_GATEWAY_API_KEY" \3 -H "Content-Type: application/json" \4 -d '{5 "model": "cerebras/gpt-oss-120b",6 "messages": [{"role": "user", "content": "Explain how Cerebras inference works"}]7 }'
Try Cerebras models in the Playground π
Get started now π