curl --request POST \
--url https://api.openai.com/v1/evals/{eval_id} \
--header 'Authorization: Bearer <token>' \
--header 'Content-Type: application/json' \
--data '
{
"name": "<string>",
"metadata": {}
}
'{
"object": "eval",
"id": "<string>",
"name": "Chatbot effectiveness Evaluation",
"data_source_config": {
"type": "custom",
"schema": "{\n \"type\": \"object\",\n \"properties\": {\n \"item\": {\n \"type\": \"object\",\n \"properties\": {\n \"label\": {\"type\": \"string\"},\n },\n \"required\": [\"label\"]\n }\n },\n \"required\": [\"item\"]\n}\n"
},
"testing_criteria": "eval",
"created_at": 123,
"metadata": {},
"share_with_openai": true
}Update certain properties of an evaluation.
curl --request POST \
--url https://api.openai.com/v1/evals/{eval_id} \
--header 'Authorization: Bearer <token>' \
--header 'Content-Type: application/json' \
--data '
{
"name": "<string>",
"metadata": {}
}
'{
"object": "eval",
"id": "<string>",
"name": "Chatbot effectiveness Evaluation",
"data_source_config": {
"type": "custom",
"schema": "{\n \"type\": \"object\",\n \"properties\": {\n \"item\": {\n \"type\": \"object\",\n \"properties\": {\n \"label\": {\"type\": \"string\"},\n },\n \"required\": [\"label\"]\n }\n },\n \"required\": [\"item\"]\n}\n"
},
"testing_criteria": "eval",
"created_at": 123,
"metadata": {},
"share_with_openai": true
}Bearer authentication header of the form Bearer <token>, where <token> is your auth token.
The ID of the evaluation to update.
Request to update an evaluation
Rename the evaluation.
Set of 16 key-value pairs that can be attached to an object. This can be useful for storing additional information about the object in a structured format, and querying for objects via API or the dashboard.
Keys are strings with a maximum length of 64 characters. Values are strings with a maximum length of 512 characters.
Show child attributes
The updated evaluation
An Eval object with a data source config and testing criteria. An Eval represents a task to be done for your LLM integration. Like:
The object type.
eval Unique identifier for the evaluation.
The name of the evaluation.
"Chatbot effectiveness Evaluation"
A CustomDataSourceConfig which specifies the schema of your item and optionally sample namespaces.
The response schema defines the shape of the data that will be:
Show child attributes
A list of testing criteria.
A LabelModelGrader object which uses a model to assign labels to each item in the evaluation.
Show child attributes
The Unix timestamp (in seconds) for when the eval was created.
Set of 16 key-value pairs that can be attached to an object. This can be useful for storing additional information about the object in a structured format, and querying for objects via API or the dashboard.
Keys are strings with a maximum length of 64 characters. Values are strings with a maximum length of 512 characters.
Show child attributes
Indicates whether the evaluation is shared with OpenAI.
Was this page helpful?