Executes an evaluator against provided input/output data for testing purposes. This endpoint allows you to test your evaluator configuration before using it in production.
All endpoints require API key authentication:
All evaluator runs now receive a single unified inputs object. This applies to all evaluator types (llm, human, code). The same fields are also recorded and visible on the Scores page for every evaluation.
New Feature: The generation_method parameter allows you to control which automation is used, since evaluators can now have both LLM and code configs.
"auto" (default): Automatically selects the best available automation method in order: LLM → Code → Legacy config"llm": Force use of LLM-based evaluation (requires llm_config to be configured)"code": Force use of code-based evaluation (requires code_config to be configured)Note: Human scoring is done through the UI/Scores API, not via this test/run endpoint.
Notes:
inputs is auto-populated from the request/response and tracing data{{llm_input}}/{{llm_output}} placeholders remain supported and transparently map to the unified fields{{input}} and {{output}}Status: 200 OK
4.5, 8.2)min_score and max_scorepassing_score thresholdtrue or false)true = passed, false = failed["Good", "Accurate"])categorical_choicesUse actual examples from your use case:
Test your evaluator configuration before production use:
API key authentication. Get your API key from https://platform.respan.ai/platform/api-keys