Back to changelog

OpenAI-Compatible Embeddings

Turn text into vectors for semantic search, clustering, and RAG — through the same gateway you already use for chat.

LLM Gateway now supports OpenAI-compatible embeddings

LLM Gateway now exposes an OpenAI-compatible /v1/embeddings endpoint. Same base URL, same API key, same SDK — point your existing OpenAI client at the gateway and embeddings.create() just works.

1curl -X POST "https://api.llmgateway.io/v1/embeddings" \2  -H "Authorization: Bearer $LLM_GATEWAY_API_KEY" \3  -H "Content-Type: application/json" \4  -d '{5    "model": "text-embedding-3-small",6    "input": "The quick brown fox jumps over the lazy dog."7  }'
  • Drop-in replacement for openai.embeddings — no code changes if you're already on the OpenAI SDK
  • Use it for semantic search, clustering, recommendations, and RAG pipelines
  • Billed on input tokens only — no output tokens, no surprises
  • Full usage and cost tracking in your dashboard, alongside your chat traffic

Browse embedding models → | Read the docs →