Human review is a critical part of evaluating AI applications. While Braintrust helps you automatically evaluate AI software with scorers, human feedback provides essential ground truth and quality assessment. Braintrust integrates human feedback from end users, subject matter experts, and product teams in one place. Use human review to:Documentation Index
Fetch the complete documentation index at: https://braintrust.dev/docs/llms.txt
Use this file to discover all available pages before exploring further.
- Evaluate and compare experiments
- Assess the efficacy of automated scoring methods
- Curate production logs into evaluation datasets
- Label categorical data and provide corrections
- Track quality trends over time
Configure review scores
Review scores let you collect structured feedback on spans and label dataset rows. only available on Pro and Enterprise plans.
Assign rows for review
You can assign rows in logs, experiments, and datasets to team members for review, analysis, or follow-up action. Assignments are particularly useful for human review workflows, where you can assign specific rows that need human evaluation and distribute review work across multiple team members. To assign a row to a team member from any table view (logs, experiments, or datasets):- Select the row.
- Select Assign.
- Choose a member to assign.
Score traces and datasets
Go to the Review page and select the type of data to review:- Log spans: production traces and debugging sessions
- Experiment spans: Evaluation results and test runs
- Dataset rows: Test cases and examples
Not all score types appear on dataset rows. Only categorical/slider scores configured to “write to expected” and free-form scores are available for dataset reviews, since datasets store test data (input/expected pairs) rather than subjective quality assessments.
Filter review data
The Review page shows any spans that have been flagged for review within a given time range. Each project provides default table views with common filters, including:- Default view: Shows all records
- Awaiting review: Shows only records flagged for review but not yet started
- Assigned to me: Shows only records assigned to you for review
- Completed: Shows only records that have finished review
scores.Preference > 0.75) to find highly-rated examples.
Change the trace layout
While reviewing log and experiment traces, you see detailed information about the flagged span by default. To switch between hierarchy, timeline, thread, and other layouts, see Examine traces.Create and edit scores inline
While reviewing, create new score types or edit existing configurations without navigating to settings:- To create a new score, click + Human review score.
- To edit an existing score, select the edit icon next to the score name.
Editing a score configuration affects how that score works going forward. Existing score values on traces remain unchanged.
Annotate in playgrounds
For a lighter-weight alternative to the full review workflow, you can annotate outputs directly in playgrounds and then get prompt improvement suggestions based on your annotations. Playground annotations help with rapid iteration during prompt development, while the Review page is better for systematic evaluation of production logs and experiments.Capture production feedback
In addition to internal reviews, capture feedback directly from production users. Production feedback helps you understand real-world performance and build datasets from actual user interactions. See Capture user feedback for implementation details and Build datasets from user feedback to learn how to turn feedback into evaluation datasets. You can also use dashboards to monitor user satisfaction trends and correlate automated scores with user feedback.Customize the review table
Show and hide columns
Select Display > Columns and then:- Show or hide columns to focus on relevant data
- Reorder columns by dragging them
- Pin important columns to the left
Use kanban layout
The kanban layout organizes flagged spans into three columns based on their review status:- Backlog: Spans flagged for review but not yet started
- Pending: Spans currently being reviewed
- Complete: Spans that have finished review
- On the Review page, select Display > Layout > Kanban.
- Drag cards between columns to update review status. Changes save automatically.
- Click any card to open the full trace for detailed review.
Create custom table views
To create or update a custom table view:- Apply the filters and display settings you want.
- Open the menu and select Save view… or Save view as….
Custom table views are visible to all project members. Creating or editing a table view requires the Update project permission.
Set default table views
You can set default views at two levels:- Organization default: Visible to all members when they open the page. This applies per page — for example, you can set separate organization defaults for Logs, Experiments, and Review. To set an organization default, you need the Manage settings organization permission (included by default in the Owner role). See Access control for details.
- Personal default: Overrides the organization default for you only. Personal defaults are stored in your browser, so they do not carry over across devices or browsers.
- Switch to the view you want by selecting it from the menu.
- Open the menu again and hover over the currently selected view to reveal its submenu.
- Choose Set as personal default view or Set as organization default view.
- Open the menu and hover over the currently selected view to reveal its submenu.
- Choose Clear personal default view or Clear organization default view.
Next steps
- Add labels and corrections to categorize and tag traces
- Build datasets from reviewed logs
- Capture user feedback from production
- Run evaluations with human-reviewed datasets