CanopyWave: 3 New Models with 75% Off
CanopyWave brings Qwen3 Coder, MiniMax M2, and GLM-4.6 to LLM Gateway with an exclusive 75% discount on all three models.

We're excited to announce three new models from CanopyWave with an exclusive 75% discount on all models!
🎯 New Models Available
GLM-4.6 - Advanced Model with Reasoning
Model ID: canopywave/glm-4.6
Context Window: 202,752 tokens
Pricing: $0.45 $0.11 per 1M input tokens / $1.50 $0.38 per 1M output tokens (75% off)
Enhanced reasoning and tool calling capabilities
Qwen3 Coder - Specialized Coding Model
Model ID: canopywave/qwen3-coder
Context Window: 262,000 tokens
Pricing:$0.22 $0.06 per 1M input tokens / $0.95 $0.24 per 1M output tokens (75% off)
Advanced coding capabilities with massive context window
MiniMax M2 - High-Performance Chat
Model ID: canopywave/minimax-m2
Context Window: 196,608 tokens
Pricing: $0.25 $0.06 per 1M input tokens / $1.00 $0.25 per 1M output tokens (75% off)
Powerful conversational AI with large context support
🚀 Getting Started
All models support streaming, tool calling, and JSON output mode:
curl -X POST https://api.llmgateway.io/v1/chat/completions \-H "Authorization: Bearer $LLM_GATEWAY_API_KEY" \-H "Content-Type: application/json" \-d '{"model": "canopywave/qwen3-coder","messages": [{"role": "user", "content": "Write a Python function"}]}'
curl -X POST https://api.llmgateway.io/v1/chat/completions \-H "Authorization: Bearer $LLM_GATEWAY_API_KEY" \-H "Content-Type: application/json" \-d '{"model": "canopywave/qwen3-coder","messages": [{"role": "user", "content": "Write a Python function"}]}'
✅ 75% Discount - Exclusive pricing for all three models
✅ Large Context Windows - 196k-262k tokens
✅ Full Feature Support - Streaming, tools, JSON output
✅ Instant Access - Available now
Get started now 🚀