InferX Beta Serverless GPU Inference Platform, Built for Agent-Native Workloads

Endpoints

Instantly usable serverless inference endpoints with OpenAI-compatible APIs for OpenCode, KiloCode, Dify, OpenWebUI, and other agent-native workloads. Built for fast integration and subsecond cold start workflows.

Browse published endpoints here. Log in to get a tenant-scoped endpoint URL and inference API key.

Browse the published endpoints first, then log in when you want a tenant-scoped URL, API key, and runnable integration flow.

1. Choose an endpoint Filter by provider, tag, or use case, then open the endpoint that matches your workload.
2. Copy setup values Use the API base URL, model name, and API key to configure Dify, OpenWebUI, Continue, OpenCode, or plain curl.
3. Validate the integration Open the endpoint detail page to copy a sample REST call or run the playground before wiring it into your app.
Public pages show the published endpoint metadata and model identifiers. Sign in to get the real tenant path and API key.
Name Brief Intro GPU Count Context Length Concurrency Integration
No endpoints are available right now.