You can update a prompt version by sending a PATCH request to the prompt version endpoint. The request should include the name of the prompt version, and the list of fallback models, load balance models, tools, and whether to deploy this version as the live version.
messages array required: The list of messages for the prompt version. If you want to add a variable, you can use the following format: {{variable_name}}.Example
model string required: Speciy the model you want to use in this version.
description string: Description of the prompt version
stream boolean: Whether the prompt version should be streamed or not. Default is false.
temperature float: The temperature of the model.
max_tokens integer: The maximum number of tokens to generate.
top_p float: The nucleus sampling probability.
frequency_penalty float: Specify how much to penalize new tokens based on their existing frequency in the text so far. Decreases the model’s likelihood of repeating the same line verbatim
presence_penalty float: Specify how much to penalize new tokens based on whether they appear in the text so far. Increases the model’s likelihood of talking about new topics
variables object: The list of variables for the prompt version. You can use these variables in the messages.
Example
fallback_models array: The list of fallback models for the prompt version. Check out fallback models for more information.Example
load_balance_models array: The list of models to load balance the prompt version. Check out load balancing for more information.Example
tools array: The list of tools to use for the prompt version. Check out tools for more information.Example
deploy boolean: Whether to deploy this version as the live version. Default is false. A newer version must exist to deploy a previous version. For example, to deploy v2, v3 must already exist.API key authentication. Get your API key from https://platform.respan.ai/platform/api-keys