Calibrate
Start Job
Starts a Scoring Spec Calibration job
POST
Authorizations
Body
application/json
Rated examples to use when calibrating the scoring spec. Must specify either the examples or the preference examples
An labeled example for training or evaluation
Example:
[
{
"llm_input": "good input",
"llm_output": "good response",
"rating": "Strongly Agree"
},
{
"llm_input": "neutral input",
"llm_output": "neutral response",
"rating": "Neutral"
}
]
Preference examples to use when calibrating the scoring spec. Must specify either the examples or preference examples
An preference example for training or evaluation
Example:
[
{
"chosen": "chosen response",
"llm_input": "some input",
"rejected": "rejected response"
}
]
Either a scoring spec or a list of questions to score
Example:
[
{
"is_lower_score_desirable": false,
"question": "Is this response truthful?"
},
{
"is_lower_score_desirable": false,
"question": "Is this response relevant?"
}
]
The strategy to use to calibrate the scoring spec. FULL would take longer than LITE but may result in better result.
Available options:
LITE
, FULL
Response
200
application/json
Successful Response
Detailed status of the job
Example:
["Downloading model", "Tuning prompt"]
The job id
Example:
"1234abcd"
Current state of the job
Available options:
QUEUED
, RUNNING
, DONE
, ERROR
, CANCELLED
The calibrated scoring spec