Create a model response using the OpenAI Responses API through the Respan gateway. Supports text and image inputs, text outputs, and stateful interactions.
This endpoint works exclusively with OpenAI models and cannot be used with models from other providers.
How to pass Respan parameters
All Respan parameters are supported. Pass them via the X-Data-Respan-Params header (base64-encoded JSON), since the Responses API body schema does not support extra top-level fields.
With prompt management
Load a managed prompt and pass runtime variables:
Streaming
Supported features
All standard OpenAI Responses API parameters are passed through to OpenAI. Additionally, Respan provides:
- Observability: All requests are logged with full telemetry (latency, tokens, cost)
- Customer tracking: Tag requests with
customer_identifier and metadata
- Prompt management: Load and render managed prompts via the
prompt parameter
- Thread grouping: Group related requests with
thread_identifier
For the full list of Respan parameters, see Chat Completions Respan Parameters.