Batch API
This endpoint creates a batch inference job with the specified model and batches.
The response contains a batch_id which can be used to track the status of the batch job.
Authorizations
x-api-keystringRequired
Body
modelstring · uuidRequiredExample:
The UUID of the model to be used for the inference job.
fd2ecd75-2e7a-4758-9613-b39a274e4f10Responses
200
Successful response with batch job details.
application/json
400
Unauthorized request.
application/json
502
Server error during inference job creation.
application/json
post
/api/v0/deploy/create-batch-inferenceThis endpoint checks the status of a specific batch inference job using the provided batch_id.
It returns the current status and, if completed, the results of the inference job.
Authorizations
x-api-keystringRequired
Body
batch_idstring · uuidRequiredExample:
The UUID of the batch job to check.
71123c09-adca-4d33-b93d-b36780e62bfbResponses
200
Successful response with batch job status and results if completed.
application/json
400
Invalid request or missing batch_id.
application/json
502
Server error during batch status check.
application/json
post
/api/v0/deploy/check-batch-inferenceLast updated