UXit Documentation
Analytics

Reading Results

Interpret score changes, weak categories, and trends across evaluation versions.

Overview

Reading results is about turning charts and scores into decisions.

The Data Visualization page shows the numbers and visuals. This page explains how to interpret them when you are reviewing a single evaluation or comparing multiple evaluations of the same flow over time.

What To Look For

When you review analytics, focus on a few simple questions:

  • Which categories are strongest right now?
  • Which categories are weakest right now?
  • Are scores improving, declining, or staying flat over time?
  • Which specific conditions caused a lower result?
  • Did the latest round of changes improve the flow in a meaningful way?

Start with the summary in Overview, then use Condition Details to confirm what caused the result.

Interpreting A Single Evaluation

If you are looking at one completed evaluation, use the results to understand current performance.

  • A higher overall score suggests the flow is meeting more of the selected criteria.
  • A low category score usually points to a concentrated problem area.
  • A balanced category pattern often means the flow is performing consistently across the guideline set.
  • A sharp drop in one category may indicate a specific weakness that deserves immediate review.

Once you identify a weak category, open Condition Details to see which individual questions failed.

Trend data becomes more useful when the same flow has been evaluated more than once against the same guideline set.

  • Improving trends usually mean recent design changes are helping.
  • Flat trends may mean the flow is stable, already optimized, or not improving in the area you expected.
  • Declining trends can point to regressions introduced by recent updates.
  • Mixed category movement can show that one improvement helped one area while creating problems in another.

Trends are most trustworthy when the flow stays the same and the guideline set stays consistent across versions.

Worked Example: Checkout Flow Analytics

Suppose you're tracking your checkout flow over three iterations against your ecommerce guidelines:

Version 1: Starting Point (D Grade)

Overall Score: 65%
Pass: 13 / 20
Categories:
  Accessibility: 60%
  Performance: 70%
  Security: 75%
  Usability: 50% <- Problem area

Insight: Usability is your biggest weakness at 50%. You're strong in Security (75%) and Performance (70%).

Version 2: Usability Focus (B Grade)

After redesigning the checkout form for clarity and adding better error messages, plus improving color contrast and ARIA labels:

Overall Score: 78%
Pass: 16 / 20
Categories:
  Accessibility: 80% ↑ +20
  Performance: 70%
  Security: 75%
  Usability: 85% ↑ +35

Insight: Your targeted fixes worked. Usability jumped 35 points and Accessibility jumped 20 points. You moved from D to B grade. Performance and Security stayed stable, confirming they do not need work yet.

Version 3: Performance Focus (B+ Grade)

Focused on optimizing images, reducing JavaScript, and implementing lazy loading:

Overall Score: 88%
Pass: 18 / 20
Categories:
  Accessibility: 80%
  Performance: 95% ↑ +25
  Security: 75%
  Usability: 85%

Insight: Performance jumped 25 points. Overall score is now 88% (B grade). The two remaining failures are edge cases. This kind of history makes it easier to connect design changes to measurable results over time.

How To Use Results Well

  • Use weak categories to decide where to investigate first.
  • Use Condition Details to identify the exact failed conditions behind a score.
  • Add notes and screenshots when a failure needs explanation or follow-up.
  • Compare evaluation versions to confirm whether a design change actually improved the flow.
  • Treat analytics as a decision-making aid, not just a reporting view.

On this page