Quarter 4 2025: Major Platform Updates

Three months of updates: 15+ new models, team management, referral program, tiered pricing, data retention, new providers, and much more.

Q4 2025 Feature Roundup

Here's everything new in LLM Gateway over the past three months — from cutting-edge models to powerful platform features.

🤖 New Models

New models

Explore models

👥 Team Management

Full team collaboration is now available:

  • Roles: Owner, Admin, and Developer with granular permissions
  • Seat-based pricing: 3 included seats, additional seats at $10/seat/month
  • Prorated billing: Fair charges based on your billing cycle
  • Enterprise options: SSO/SAML, SCIM provisioning, audit logs for larger teams

Learn more | Enterprise

🏢 Admin Dashboard

New internal admin dashboard for platform management with:

  • Token metrics calculation and aggregation
  • User and organization management
  • Performance monitoring

💰 Referral Program

Earn 1% of every referral's spending — forever!

  • Automatic tracking via referral links and cookies
  • Earnings credited to your account
  • No cap on earnings

Find your referral link in Settings → Referrals.

Learn more here

🛡️ Routing Improvements

Routing Improvements

  • Uptime-aware routing — Automatically routes away from low-uptime providers
  • X-No-Fallback header — Prevent automatic provider fallback when needed
  • Provider metrics — Routing decisions include provider performance data
  • Multi-API key support — Load balance across multiple provider keys
  • Latency optimization — Only consider latency when streaming is enabled

Learn more

📊 Tiered Pricing

Volume-based pricing is now available:

  • The more you use, the more you save
  • Tiered cache pricing for supported models
  • Discounts calculated automatically based on usage
  • View pricing tiers on any model's detail page

🔒 Data Retention & Storage

Data Retention

New data retention controls with billing:

  • Metadata-only (default): Only request metadata is stored
  • Full retention: Store complete request and response data
  • Data storage billing: Pay only for what you store
  • Data limits: Set maximum storage limits per organization

Configure in Settings → Organization → Data Retention.

🔍 Model Discovery

Search bar in the top navigation — instantly find any model across all providers.

Model Timeline & Metadata

  • Release dates (releasedAt, publishedAt) for all models
  • Model descriptions on detail pages
  • Visual indicators for free models and provider mappings
  • Filtered counts showing matching models and providers

Explore timeline

⚖️ Model Comparison

Compare models side-by-side with our new comparison page:

  • Input/output token pricing — Compare costs across providers
  • Feature comparison — See capabilities like streaming, vision, tools at a glance
  • Context windows — Compare model limits instantly

Compare models

🎮 Playground Enhancements

Playground

  • Comparison mode — Compare responses from multiple models side-by-side
  • Group chat — Chat with multiple models simultaneously
  • GitHub MCP tools — Integrated GitHub operations via Model Context Protocol
  • Redirect after signin — Return to your original URL after authentication
  • Image configuration — Set image parameters and reasoning effort options

Try out Chat Playground

📈 Analytics & Metrics

New dashboard

  • Cost breakdown — Detailed breakdown in usage API responses
  • Discount savings visualization — See savings on dashboard
  • Throughput metrics — Monitor tokens per second
  • Provider tracking — Track which providers serve requests
  • Real-time metrics — Current minute history calculation
  • Reasoning tokens — Included in cost calculations

Learn more

💳 Billing Enhancements

Billing

  • PDF invoices — Automatically generated and emailed for subscriptions and credits
  • Company address on invoices
  • Billing email management — Set separate billing contact
  • Top-up credits button — Quick access to add credits
  • Refund tracking — Full visibility in transaction history
  • First-time bonus — Bonus credits on first purchase

🔐 Security Enhancements

  • Email verification required — New accounts must verify before API access
  • Enhanced email validation — Blocks plus signs, disposable emails, blacklisted domains
  • Masked provider keys — Tokens masked in API responses
  • Content filter detection — Proper error handling for filtered content

💸 Discounts

  • 20% off all Google models — Applied automatically
  • 10% off all Z.ai models — Applied automatically
  • 75% off CanopyWave models — Qwen3-Coder, MiniMax M2, GLM-4.6, Kimi K2 Thinking
  • 90% off DeepSeek V3.1 via CanopyWave partnership

View discounted models

🌐 New Provider Support

We've added support for major cloud providers and new partners:

  • AWS Bedrock — Access Llama and other models via Amazon's managed service
  • Microsoft Azure — Native Azure OpenAI integration
  • Google Vertex AI — Full Vertex AI support with global routing
  • CanopyWave — Discounted access to DeepSeek, Kimi, and more
  • NanoGPT — New provider option

Browse all providers

📱 API Enhancements

  • Effort parameter — Control reasoning depth for Claude Opus 4.5
  • no_reasoning flag — Disable reasoning when not needed
  • Responses API — Support for GPT-5 Pro and GPT-5.1 Codex
  • Thought signatures — Included in Google tool calls
  • Custom upload limits — Configure per-gateway limits
  • Unsupported message handling — Clear feedback for unsupported features
  • JSON schema response format — Structured output with json_schema response format
  • JSON Output filter — Filter models by JSON output support in UI
  • OpenAI JSON schema conversion — Automatic format conversion for Google models
  • DeepSeek JSON schema — Native JSON schema output support for DeepSeek V3.1

📚 Documentation

  • Cline integration guide — Use LLM Gateway with Cline
  • Cursor guide — IDE integration documentation
  • Data retention & caching docs — Full documentation
  • Reasoning docs — Guidelines for reasoning models
  • Collapsible links — Improved navigation
  • GitHub timestamps — Last edit dates on all pages

Visit our Docs

📧 Email & Notifications

  • Transactional emails — Trial start and cancellation notifications
  • PDF invoice emails — Automatic invoice delivery
  • Onboarding status — Synced with Brevo for marketing

🎉 Other Improvements

  • Organization name editing — Update org name in preferences
  • Free model rate limits — Tiered rate limiting for free models
  • Sensitive word check — Parameter support for Z.ai
  • Decimal.js — Precise cost calculations
  • Shared UI package — Reusable components across apps

📢 Pro Plan Pricing Update

Starting December 19th, 2025, Pro plan fees will increase from 0% to 1%. This remains 80% less than the 5% fee on the Free plan, ensuring you continue to get exceptional value as you scale.


Explore all models | Try the Playground | Get started now 🚀

    Quarter 4 2025: Major Platform Updates - Blog - LLM Gateway