Starting with version 2408.05 of PIX on Windows, the Timing Capture comparison features have been moved from the Metrics view to a new dedicated Comparison Layout. This layout improves several comparison workflows, including the ability to compare data from multiple captures.
The Comparison Layout produces a statistical comparison of the average duration for the points that represent PIX event hierarchies in two selected ranges of time, or for the points above and below a metric’s budget. Â Statistical comparisons help determine which portions of the event hierarchies had statistically different durations for the set of points being compared.
The layout consists of a metric selection panel, graphs for the left and right set of points to be compared, a table showing the results of the comparison, and a histogram of the values of both the left and right sets of points for the selected row in the table. Â The set of points for the right side of the comparison may come from a different capture than the points from the left side of the comparison. Â This allows you to compare data from a capture taken on one type of Windows device with the data taken on a different device, for example.
Comparing groups of points for a metric
To produce a comparison for a metric, select Metrics from the Comparison Item dropdown, select a metric from the selection panel or from the Recent Metrics dropdown, then select a range of time in both the left and right graphs. Â For example, the following graph of a metric named Game::Render shows several ranges of time where the event durations are relatively consistent, but there are also ranges of time in which the event’s durations are significantly different between the baseline and experiment graphs. Â A statistical comparison of the event durations in the two highlighted time ranges can be used to determine which portions of the Game::Render event hierarchy account for the differences in duration.
Note that each selected range of time must have at least two points for the metric being compared. Â It’s recommended that you select ranges that have enough points to provide a good sample representation of your game’s behavior. The amount of data may vary depending on how noisy your source data is, but a good rule of thumb is to look for ranges with 50-100 data points. A warning indicator will be shown if the number of data points is below 20.
The Left column in the table displays the average duration for the set of points in the range of time selected in the left graph. Â The Right column displays the average duration for the set of points in the range of time selected in the right graph. The Delta and Delta % columns show the differences in duration between the left and right groups.
The P-value column identifies whether the differences between the left and right groups are statistically significant or whether the differences are due to other factors such as noise, random sampling, or sampling bias. Â The lower the P-value, the more statistically significant the differences are. Â Those rows with the lowest P-values, are the rows you’ll likely want to investigate further by drilling further into the event hierarchy. Â Rows that are statistically significant are colored in either red or green depending on whether the delta between the baseline and experiment groups is negative or positive. Â The boldness of the red and green colors is relative to how close to 0 the P-value is. Â Up and down arrows are also displayed in the P-value column to aid with accessibility. Â Sorting by P-value is a convenient way to bring the most statistically significant points to the top of the table.
The histogram at the bottom of the layout shows the distribution of the durations for the left and right groups for the selected row in the table. Â There is generally a correlation between the P-value for a given row and its histogram. Â The histograms for the left and right groups are largely overlapping for rows with high P-values, and vice versa.
The number of points in the left and right groups also influences how to interpret the results. Â The N columns in the table display the number of points for each group. Â A yellow warning triangle is drawn in those cells with low sample counts. Â Results may be biased in these cases.
Comparing the points above and below a metric’s budget
To compare the points above and below a metric’s budget, first create a metrics view configuration that includes the metrics and the budgets to analyze. Â Then, select Budgets from the Comparison Item dropdown, the desired metrics view configuration from the Configuration dropdown, and the budget to analyze from the Budget dropdown.
When comparing the points above and below a budget, an Above/Below slider appears above both the left and right graphs. Ensure that Above is selected for one graph, and Below is selected from the other.
Note that any metric can be added to the comparison using the Selector panel. The list of metrics is not limited to only those that have budgets defined. This enables you to look at everything that is happening in a capture when a single metric is out of budget and compare it to what happens in the capture when the metric is within budget.
The comparison layout can also be accessed from the metrics view. Â Select the calculator icon next to the budget to open the comparison layout.
As always, please send us bug reports and feature requests using the PIX Feedback button in the upper right of the UI.
Steven.
0 comments
Be the first to start the discussion.