Create a prompt version
Create a new version of a prompt. Use {{variable_name}} syntax in messages to define template variables.
Path parameters
prompt_id
The unique prompt identifier.
Headers
Authorization
Bearer token. Use Bearer YOUR_API_KEY.
Request
This endpoint expects an object.
messages
Messages for this version. Use {{variable_name}} placeholders for template variables.
model
Primary model for this version.
description
Version description.
thinking
Optional provider-specific reasoning configuration.
stream
Whether to stream responses.
temperature
Sampling temperature (0-2).
max_tokens
Maximum tokens to generate.
top_p
Nucleus sampling parameter.
frequency_penalty
Frequency penalty (-2 to 2).
presence_penalty
Presence penalty (-2 to 2).
reasoning_effort
verbosity
seed
variables
Template variables and their default values.
fallback_models
Fallback models if the primary model fails.
load_balance_models
Weighted load-balancing model configuration.
tools
Tools available to the model.
tool_choice
response_format
Structured output / response format configuration.
json_schema
JSON schema used for structured outputs when configured.
is_enforcing_response_format
Whether to strictly enforce the response format.
deploy
Deploy this version as the live version immediately.
Response
Prompt version created successfully.
id
prompt_version_id
version
description
messages
thinking
model
stream
temperature
max_tokens
top_p
frequency_penalty
presence_penalty
reasoning_effort
verbosity
seed
variables
fallback_models
load_balance_models
tools
tool_choice
response_format
json_schema
is_enforcing_response_format
readonly
is_deployed
edited_by
created_at
updated_at
Errors
400
Bad Request Error
401
Unauthorized Error
404
Not Found Error