Start with the work
Bring the prompt, model output, risk level, and task type into one authoring surface.
Turn prompts, model outputs, and expert judgment into governed rubrics, graded evaluations, and model-performance memory.
What it is
Rubric Studio gives teams a governed way to define how model outputs are judged. It keeps authoring, AI-drafted criteria, expert approval, worker grading, evidence, QA, adjudication, scorecards, and exports attached to the same evaluation record.
How it works
Bring the prompt, model output, risk level, and task type into one authoring surface.
Write criteria by hand or let AI draft the first spine for expert review.
AI-drafted rubrics stay blocked until an expert approves and activates them.
Workers score each criterion, attach required evidence, and see blockers before submit.
Submitted grades contribute to model scorecards and regression memory.
Who it is for
Roles
Criteria can require quotes, source citations, screenshots, test output, or reviewer notes. Submit stays blocked until required criteria and required evidence are complete.