Endpoints
Instantly usable serverless inference endpoints with OpenAI-compatible APIs for OpenCode, KiloCode, Dify, OpenWebUI, and other agent-native workloads. Built for fast integration and subsecond cold start workflows.
Browse published endpoints here. Log in to get a tenant-scoped endpoint URL and inference API key.
Browse the published endpoints first, then log in when you want a tenant-scoped URL, API key, and runnable integration flow.
1. Choose an endpoint
Filter by provider, tag, or use case, then open the endpoint that matches your workload.
2. Copy setup values
Use the API base URL, model name, and API key to configure Dify, OpenWebUI, Continue, OpenCode, or plain curl.
3. Validate the integration
Open the endpoint detail page to copy a sample REST call or run the playground before wiring it into your app.
Public pages show the published endpoint metadata and model identifiers. Sign in to get the real tenant path and API key.
| Name | Brief Intro | GPU Count | Context Length | Concurrency | Integration |
|---|---|---|---|---|---|
| No endpoints are available right now. | |||||