Video Generation, Sessions, 3-Mode Content Filter & More
Generate videos via the API, track conversations with sessions, filter content with a new 3-mode content filter, and more — plus new models and providers.

Video Generation
LLM Gateway now supports video generation via the API. Generate videos from text prompts using supported models, with job-based async processing and status tracking.
- Submit a video generation request and receive a job ID
- Poll for job status — completed jobs return the video URL
- Failed jobs are properly marked with error details
Read the video generation docs
Sessions
Sessions let you track and organize conversations across multiple requests:
- Group related requests into named sessions
- Filter activity logs by session
- View streaming cost breakdowns per session
- Improved empty state UI when no sessions exist
3-Mode Content Filter
A new 3-mode content filter gives you granular control over content moderation at the gateway level:
- Off — No filtering, all requests pass through
- Flag — Requests are flagged but still processed
- Block — Flagged requests are blocked before reaching the provider
Configure the filter per project to match your safety requirements.
New Models
GPT-5.4 Mini & Nano
1openai/gpt-5.4-mini2openai/gpt-5.4-nano
1openai/gpt-5.4-mini2openai/gpt-5.4-nano
OpenAI's smaller GPT-5.4 variants are now available — offering the same architecture at lower cost for lighter workloads.
View GPT-5.4 Mini | View GPT-5.4 Nano
MiniMax M2.7
1minimax/minimax-m2.7
1minimax/minimax-m2.7
New MiniMax M2.7 model mappings added across providers.
New Provider: EmberCloud
EmberCloud is now available as a provider, expanding your routing options with additional model coverage.
UI & Platform Improvements
- Redesigned Code app — Fresh look for the dev plans and coding tools dashboard
- Revamped admin dashboard — Improved performance on models, mappings, and providers pages
- Cost simulator revamp — Updated cost simulator with better navigation
- Activity log filters — Improved filtering in activity logs
- Status page indicator — Quick link to the status page from the dashboard
- OpenCode support — Added OpenCode to the integrations list
- Free icon visibility — Free icon now correctly hidden for paid provider mappings
Documentation & Guides
- Codex CLI guide — Step-by-step guide for using LLM Gateway with OpenAI Codex CLI
- Autohand guide — Integration guide for Autohand
- OpenClaw guide — Integration guide for OpenClaw
- Introduction page — New docs introduction page
- Themed screenshots — Light/dark themed screenshots in the learn section
Fixes & Reliability
- Moonshot thinking disabled for tool calls to prevent errors
- ByteDance sensitive content blocks now properly classified
- Cached tokens no longer double-counted in storage cost calculations
- Bedrock cached tokens correctly included in streaming responses
- Streaming error diagnostics improved with better logger serialization
- Canopywave marked as unstable due to availability issues
- Top-tier model provider compatibility fixes
- XAI reasoning output option added to configuration