Hosted endpoint
Your app sends requests to our hosted /v1/chat/completions endpoint using your OpenAI API key as the bearer token.
Docs
These docs are intentionally lightweight. They explain the product contract users actually need: your own OpenAI key, our hosted /v1, private signed dashboard access, model presets, and tier roles.
Your app sends requests to our hosted /v1/chat/completions endpoint using your OpenAI API key as the bearer token.
Each API key maps to a tenant-scoped model policy. Users can change their own utility, fast, and deep tiers from the signed settings page.
Every tenant gets a signed private dashboard URL from the response header. No separate end-user auth flow is required for the client experience.
Pushes the most traffic into the lowest-cost path. Best when speed and spend matter more than richer answers.
The default preset. Good for most teams that want reliable output quality and obvious savings without extra tuning.
Uses stronger fast and deep tiers for teams prioritizing answer quality over token savings.
Lowest-cost work. Good for small transforms, boilerplate, and routine low-stakes traffic.
Main production tier. This should handle most traffic well when the prompt does not need extra depth.
Highest-cost tier. Reserve it for harder prompts, richer explanation, or cases where completeness matters.
You use your own OpenAI API key. This product is a routing layer and hosted endpoint, not a separate billing identity.
Point your client to our hosted base URL and send OpenAI-compatible traffic to /v1/chat/completions.
Read the X-Client-Dashboard-Url response header from your first successful request and open that signed link.
Yes. Open /client/settings from your signed dashboard flow and choose a preset or set utility, fast, and deep models directly.
No. The client pages are intentionally separate from admin. End users only get their own private dashboard and settings flow.