Annotations are now visible directly on the Experiment Compare page, making it easier to review context and insights alongside your experiment results. This streamlines analysis by keeping annotations and metrics in one place.
Annotations on the Experiment Compare page now autosave immediately after selection, removing the need to manually save them. Experiment annotations are also displayed on the experiment compare page.
Dashboards now support per-widget time range selection and improved legend display options, giving users more granular control and clearer data visualization. These updates make it easier to tailor each dashboard view to your specific analysis needs.
The Playground now supports running on an unlimited number of rows, with a new “Load 100 more” button to make adding records easier. This is the first iteration of scalable runs, and we’ll continue improving the usability in upcoming updates.
Session- and trace-level evaluations are available across all Arize tiers. This unlocks the ability to evaluate at different levels (span, trace, and session), which is crucial for building and monitoring agents.
You can now generate synthetic datasets directly in Playground using Alyx. This makes it easier to quickly create sample data for testing and experimentation without needing external tools.
Added support for defining optimization direction on categorical and continuous configs. The optimization_direction field now lives within their respective config columns, improving logical grouping.
Improved Annotator Selection for New Labeling Queues
The annotator selection flow has been refined: users are grouped into Annotators and All Other Users (with group-level select all), and the interface now clearly highlights “You.”
Dashboards have been reworked with a richer experience: clickable legends and new widget creation forms for line charts, bar charts, experiments, monitors, and statistics.