Health

Server Live

get

The “server live” API indicates if the inference server is able to receive and respond to metadata and inference requests. The “server live” API can be used directly to implement the Kubernetes livenessProbe.

Responses
200

OK

get
GET /v2/health/live HTTP/1.1
Host: 
Accept: */*
200

OK

No content

Server Ready

get

The “server ready” health API indicates if all the models are ready for inferencing. The “server ready” health API can be used directly to implement the Kubernetes readinessProbe.

Responses
200

OK

get
GET /v2/health/ready HTTP/1.1
Host: 
Accept: */*
200

OK

No content

Model Ready

get

The “model ready” health API indicates if a specific model is ready for inferencing. The model name and (optionally) version must be available in the URL. If a version is not provided the server may choose a version based on its own policies.

Path parameters
model_namestringRequired
model_versionstringRequired
Responses
200

OK

get
GET /v2/models/{model_name}/versions/{model_version}/ready HTTP/1.1
Host: 
Accept: */*
200

OK

No content

Model Ready

get

The “model ready” health API indicates if a specific model is ready for inferencing. The model name and (optionally) version must be available in the URL. If a version is not provided the server may choose a version based on its own policies.

Path parameters
model_namestringRequired
Responses
200

OK

get
GET /v2/models/{model_name}/ready HTTP/1.1
Host: 
Accept: */*
200

OK

No content

Was this helpful?