Update an Evaluation.
Update the setup of an Evaluation by specifying the Dataset, versions to be evaluated (Evaluatees), and which Evaluators to provide judgments.
Unique identifier for Evaluation.
Dataset to use in this Evaluation.
Unique identifiers for the Prompt/Tool Versions to include in the Evaluation. Can be left unpopulated if you wish to add evaluatees to this Evaluation by specifying evaluation_id
in Log calls.
The Evaluators used to evaluate.
Name of the Evaluation to help identify it. Must be unique within the associated File.
The File to associate with the Evaluation.
Unique identifier for the Evaluation. Starts with evr
.
The Dataset used in the Evaluation.
The Prompt/Tool Versions included in the Evaluation.
The Evaluator Versions used to evaluate.
The current status of the Evaluation.
"pending"
: The Evaluation has been created but is not actively being worked on by Humanloop."running"
: Humanloop is checking for any missing Logs and Evaluator Logs, and will generate them where appropriate."completed"
: All Logs an Evaluator Logs have been generated."cancelled"
: The Evaluation has been cancelled by the user. Humanloop will stop generating Logs and Evaluator Logs.Name of the Evaluation to help identify it. Must be unique among Evaluations associated with File.
Unique identifier for the File associated with the Evaluation.
URL to view the Evaluation on the Humanloop.