- Evaluate and compare experiments
- Assess the efficacy of automated scoring methods
- Curate production logs into evaluation datasets
- Label categorical data and provide corrections
- Track quality trends over time
Configure review scores
Review scores let you collect structured feedback on spans and label dataset rows. Configure scores in Settings > Project > Human review. See Configure human review for details on score types and options.Assign rows for review
You can assign rows in logs, experiments, and datasets to team members for review, analysis, or follow-up action. Assignments are particularly useful for human review workflows, where you can assign specific rows that need human evaluation and distribute review work across multiple team members. To assign a row to a team member from any table view (logs, experiments, or datasets):- Select the row.
- Select Assign.
- Choose a member to assign.
Score traces and datasets
Go to the Review page and select the type of data to review:- Log spans: production traces and debugging sessions
- Experiment spans: Evaluation results and test runs
- Dataset rows: Test cases and examples
Not all score types appear on dataset rows. Only categorical/slider scores configured to “write to expected” and free-form scores are available for dataset reviews, since datasets store test data (input/expected pairs) rather than subjective quality assessments.
Filter review data
The Review page shows any spans that have been flagged for review within a given time range. Each project provides default table views with common filters, including:- Default view: Shows all records
- Awaiting review: Shows only records flagged for review but not yet started
- Assigned to me: Shows only records assigned to you for review
- Completed: Shows only records that have finished review
scores.Preference > 0.75) to find highly-rated examples.
Change the trace layout
While reviewing log and experiment traces, you see detailed information about the flagged span by default.View as a timeline
While viewing a trace, select Timeline to visualize the trace as a gantt chart. This view shows spans as horizontal bars where the width represents duration. Bars are color-coded by span type, making it easy to identify performance bottlenecks and understand the execution flow.View as a conversation
While viewing a trace, select Thread to view the trace as a conversation thread. This view displays messages, tool calls, and scores in chronological order, ideal for debugging LLM conversations and multi-turn interactions. By default, the thread view renders raw span data. Select to apply a preprocessor — choose the built-in Thread preprocessor to format the trace as a readable conversation, or select a custom preprocessor to control exactly how messages are rendered. When topics are enabled, topic tags and facet outputs appear at the top of the thread view as well. Use Find or pressCmd/Ctrl+F to search within the thread view and quickly locate specific content such as message text and score rationale. Matches are highlighted in-place using your browser’s native highlighting. This search is scoped to the thread view content — use the trace view’s search feature to search across spans.
Test and apply signals
While viewing a trace, select Signals to test topic facets and scorers on the current trace.- Topic facets: Test how preprocessors transform the trace data, test what summaries prompts extract, or apply the complete facet (preprocessor + prompt) to see the end-to-end result.
- Scorers: Test scorers, apply them to the trace, or configure an automation rule for online scoring.
Create custom trace views
While viewing a trace, select Views to create custom visualizations using natural language. Describe how you want to view your trace data and Loop will generate the code. For example:- “Create a view that renders a list of all tools available in this trace and their outputs”
- “Render the video url from the trace’s metadata field and show simple thumbs up/down buttons”
Self-hosted deployments: If you restrict outbound access, allowlist
https://www.braintrustsandbox.dev to enable custom views. This domain hosts the sandboxed iframe that securely renders custom view code.Search within a trace
While viewing a trace, use Find or pressCmd/Ctrl+F to search for content within the trace. A scope dropdown lets you choose where to search:
- This span — Search only within the currently selected span.
- Full trace — Search across all spans in the trace.
Trace search finds content within the currently open trace. To search across all traces in your project, use filters or deep search.
Change span data format
When viewing a trace, each span field (input, output, metadata, etc.) displays data in a specific format. Change how a field displays by selecting the view mode dropdown in the field’s header. Available views:- Pretty - Parses objects deeply and renders values as Markdown (optimized for readability)
- JSON - JSON highlighting and folding
- YAML - YAML highlighting and folding
- Tree - Hierarchical tree view for nested data structures
- LLM - Formatted AI messages and tool calls with Markdown
- LLM Raw - Unformatted AI messages and tool calls
- HTML - Rendered HTML content
Create and edit scores inline
While reviewing, create new score types or edit existing configurations without navigating to settings:- To create a new score, click + Human review score.
- To edit an existing score, select the edit icon next to the score name.
Editing a score configuration affects how that score works going forward. Existing score values on traces remain unchanged.
Annotate in playgrounds
For a lighter-weight alternative to the full review workflow, you can annotate outputs directly in playgrounds and then get prompt improvement suggestions based on your annotations. Playground annotations help with rapid iteration during prompt development, while the Review page is better for systematic evaluation of production logs and experiments.Capture production feedback
In addition to internal reviews, capture feedback directly from production users. Production feedback helps you understand real-world performance and build datasets from actual user interactions. See Capture user feedback for implementation details and Build datasets from user feedback to learn how to turn feedback into evaluation datasets. You can also use dashboards to monitor user satisfaction trends and correlate automated scores with user feedback.Customize the review table
Show and hide columns
Select Display > Columns and then:- Show or hide columns to focus on relevant data
- Reorder columns by dragging them
- Pin important columns to the left
Use kanban layout
The kanban layout organizes flagged spans into three columns based on their review status:- Backlog: Spans flagged for review but not yet started
- Pending: Spans currently being reviewed
- Complete: Spans that have finished review
- On the Review page, select Display > Layout > Kanban.
- Drag cards between columns to update review status. Changes save automatically.
- Click any card to open the full trace for detailed review.
Create custom table views
To create or update a custom table view:- Apply the filters and display settings you want.
- Open the menu and select Save view… or Save view as….
Custom table views are visible to all project members. Creating or editing a table view requires the Update project permission.
Set default table views
You can set default views at two levels:- Organization default: Visible to all members when they open the page. This applies per page — for example, you can set separate organization defaults for Logs, Experiments, and Review. To set an organization default, you need the Manage settings organization permission (included by default in the Owner role). See Access control for details.
- Personal default: Overrides the organization default for you only. Personal defaults are stored in your browser, so they do not carry over across devices or browsers.
- Switch to the view you want by selecting it from the menu.
- Open the menu again and hover over the currently selected view to reveal its submenu.
- Choose Set as personal default view or Set as organization default view.
- Open the menu and hover over the currently selected view to reveal its submenu.
- Choose Clear personal default view or Clear organization default view.
Next steps
- Add labels and corrections to categorize and tag traces
- Build datasets from reviewed logs
- Capture user feedback from production
- Run evaluations with human-reviewed datasets