Encyclopedia Evalica / Evaluation / Alignment

Alignment illustration

Alignment

/uh'leyen.muhnt/The process of calibrating an LLM judge so its scores match human judgment. Alignment typically involves iterating on rubrics and prompts until agreement is high. (noun)

After our alignment sessions, the LLM judge scores matched human ratings much more closely.

Customer example

Navan treated alignment as an iterative process, tuning its eval prompt like an ML classifier until the evaluator reached >0.9 macro F1 and using reasoning to debug failures. Read more

Related Evaluation terms

From the docs

Get started with Evals

Braintrust is the AI observability and eval platform for production AI. By connecting evals and observability in one workflow, teams at Notion, Stripe, Zapier, Vercel, and Ramp ship quality AI products at scale.

Start building