Inference API
This endpoint runs inference using a specified model deployment. It accepts optional chat history to provide context for chat-based models. The response includes the model's output and other relevant information.
Authorizations
x-api-keystringRequired
Body
modelstring · uuidRequiredExample:
The UUID of the model to be used for inference.
a4a2feb3-efc1-49d6-96ca-5f7ec05cde98deployment_versionstring | nullableOptionalExample:
The ID of the specific deployment version to use. If not provided, the default version will be used.
1.0.0Responses
200
Successful response with inference results.
application/json
400
Bad request due to missing or invalid parameters.
application/json
401
Unauthorized request due to missing or invalid API key.
application/json
502
Server error during inference process.
application/json
post
/api/v0/deploy/inferenceLast updated
Was this helpful?