Create Experiment Stream

POST

Response

This endpoint returns a stream of object.
datalist of objects
Array containing the chat responses.
provider_responseslist of any
The raw responses returned by the model provider.
project_idstringOptional
Unique identifier of the parent project. Will not be provided if the request was made without providing a project name or id
num_samplesintegerOptionalDefaults to 1
The number of chat responses.
logprobsintegerOptional

Include the log probabilities of the top n tokens in the provider_response

suffixstringOptional
The suffix that comes after a completion of inserted text. Useful for completions that act like inserts.
userstringOptional
End-user ID passed through to provider call.
usageobjectOptional
Counts of the number of tokens used and related stats.
metadatamap from strings to anyOptional
Any additional metadata to record.
provider_requestmap from strings to anyOptional
The raw request sent to the model provider.
session_idstringOptional
ID of the session if it belongs to one.
tool_choice"none" or "auto" or "required" or objectOptional
Controls how the model uses tools. The following options are supported: 'none' forces the model to not call a tool; the default when no tools are provided as part of the model config. 'auto' the model can decide to call one of the provided tools; the default when tools are provided as part of the model config. Providing {'type': 'function', 'function': {name': <TOOL_NAME>}} forces the model to use the named function.