Create an evaluation.
String ID of project. Starts with pr_
.
ID of the config to evaluate. Starts with config_
.
IDs of evaluators to run on the dataset. IDs start with evfn_
ID of the dataset to use in this evaluation. Starts with evts_
.
API keys required by each provider to make API calls. The API keys provided here are not stored by Humanloop. If not specified here, Humanloop will fall back to the key saved to your organization. Ensure you provide an API key for the provider for the model config you are evaluating, or have one saved to your organization.
Whether the log generations for this evaluation should be performed by Humanloop. If False
, the log generations should be submitted by the user via the API.
Name of the Evaluation to help identify it.
Successful Response
Unique ID for the evaluation. Starts with ev_
.
Status of an evaluation.