InferiaLLMInferiaLLM
API Reference

Orchestration Gateway

Compute and Deployment Management

The Orchestration Gateway manages the lifecycle of model deployments.

Deployments

GET /deployment/deployments

List all active deployments across all compute pools.

org_id

Querystring

Filter by Organization ID.

POST /deployment/deploy

Create a new model deployment.

model_name

BodystringRequired

Name of the model to deploy.

image

BodystringRequired

Docker image to use (e.g., vllm/vllm-openai:latest).

pool_id

BodystringRequired

Target Compute Pool ID.

replicas

Bodyinteger

Number of replicas (default: 1).

POST /deployment/terminate

Terminate a running deployment (stops the containers/service).

deployment_id

BodystringRequired

The ID of the deployment to stop.

DELETE /deployment/delete/{deployment_id}

Permanently remove a stopped or failed deployment record.

deployment_id

PathstringRequired

The ID of the deployment to delete.

Logs

GET /deployment/logs/{deployment_id}

Retrieve live logs for a specific deployment. Supports DePIN (IPFS) logs.

Model Registry

POST /deployment/registerModel

Register a new model configuration in the system.

model_name

BodystringRequired

Unique identifier for the model.

backend

BodystringRequired

Inference backend (e.g., vllm, ollama).

GET /deployment/getModel/{name}/{version}

Get details for a registered model version.

On this page