After deploying a model, use
POST /projects/:project_id/inference to run predictions through the project endpoint — no need to track training job IDs in your application code.List projects
GET /projects
Returns all projects in your account.
Array of project objects.
Create a project
POST /projects
Creates a new project.
Request body
A unique, human-readable name for the project.
The new project’s unique ID. Use this in all subsequent project and deployment requests.
The project name.
Delete a project
DELETE /projects/:project_id
Permanently deletes a project and its deployment history. Any deployed models associated with the project will no longer be accessible through the project inference endpoint.
Path parameters
The project ID.
204 No Content on success.
Deploy a model to a project
POST /projects/:project_id/deployments
Deploys a trained model to a project endpoint, making it available for inference. Creating a new deployment replaces the currently active model for that project.
Path parameters
The project ID to deploy to.
The ID of the completed training job whose model you want to deploy. The training job must be in
complete status.Unique ID for this deployment.
Current deployment status. Values:
deploying, active, failed.The training job ID that was deployed.
List deployment history
GET /projects/:project_id/deployments
Returns the full deployment history for a project, ordered from most recent to oldest. Use this to track which model versions have been deployed and when.
Path parameters
The project ID.
Run inference on a project model
POST /projects/:project_id/inference
Runs a prediction against the currently active model deployed to this project. This endpoint abstracts away the underlying training job ID so your application only needs to reference the project.
Path parameters
The project ID.
The input text to run predictions on.
The output schema for encoder models. Accepts optional keys:
entities (list of entity type strings), classifications (list of {task, labels} objects), structures, and relations. For decoder models, use "task": "generate" instead.
