GLM-4.6V Flash

Ultra-fast, lightweight vision-language model (9B) for low-latency workloads.

glm-4.6v-flash
STABLE
128,000 context
Starting at Free input tokens
Starting at Free output tokens
Streaming
Vision
Tools
Reasoning
JSON Output

Providers for GLM-4.6V Flash

LLM Gateway routes requests to the best providers that are able to handle your prompt size and parameters.

Z AI

zai/glm-4.6v-flash
Context Size
128k
Stability
STABLE
Pricing
Input
Cached
Output
Capabilities
Streaming
Vision
Tools
Reasoning
JSON Output
Try in Playground
    GLM-4.6V Flash – AI Model on LLM Gateway