Skip to content

Inference API

Use Inference API when you want programmatic access to RemoteGPU models without operating Kubernetes directly.

This product path is designed for developers, application teams, and integrators who want HTTP-based model access with API keys.

Open in the console

TaskConsole page
Review image models and runtime statusInference API / Image
Create the required API keySettings / API Keys

Choose this path when

You want to...Use Inference API if...
call models from your app or backendyou want an HTTP API instead of a hosted application
automate prompts and request parametersyou are comfortable with API keys and request payloads
avoid Kubernetes operationsyou do not want to run namespace-scoped workloads yourself

How this product path works

AreaWhat to expect
Main interfaceHTTP API, with console support for visibility
AuthenticationAPI key
Kubernetes knowledge requiredNo
Runtime modelRemoteGPU serves the model and executes requests for you

Current API guides

GuideUse it for
Image inferenceSend image-generation requests and poll job status

How this differs from the other product paths

Product pathBest forYou manage
ApplicationGuided hosted workflowsVery little beyond normal console actions
Inference APISemi-professional and programmatic useAPI calls, request payloads, and keys
KubernetesProfessional operatorsNative Kubernetes workloads and networking resources

RemoteGPU customer documentation