Evaluations
What is an Evaluation?
An evaluation is your assessment of a flow against a specific guideline set, aligned with a predefined user journey. It records which design criteria pass, fail, or are not applicable, generating a score and detailed results that you can track over time. Afterward, multiple graph types are provided to improve data visualization, or you can export the results as JSON for use in tools like MATLAB or Excel.
Before You Evaluate
Guideline Set (Required): You must have a guideline set selected. The guideline set defines all the criteria you're testing against—this is mandatory and cannot be skipped.
Flow: You need a flow to evaluate. Flows can be created from imported use cases or defined independently. Either way, the flow is the hyperscoped user journey (swimlane) you're assessing.
Creating an Evaluation
Navigate to your flow, select your guideline set, and click "Run Evaluation". You'll be presented with all questions from your chosen guideline set, organized by category.
Answering Questions
For each question, choose one of three response options:
Pass → Design meets this criterion successfully
Fail → Design does not meet this criterion
N/A → Criterion is not applicable to this evaluationYou can also add notes to each question to document your reasoning and design decisions.
Evaluation Versions & Tracking Progress
Each time you evaluate the same flow with the same guideline set, it becomes a new version. Here's how progression works:
Initial Design Assessment
↓
v1 Evaluation (65% overall score)
↓
Make Design Changes
↓
v2 Evaluation (78% overall score)
↓
Make More Improvements
↓
v3 Evaluation (85% overall score)This versioning approach builds a history that Analytics uses to show trends and measure progress.
Continue vs. Create New:
- Continue: Same flow + same guidelines = new version for trend tracking
- Create New: Different flow or guidelines = separate evaluation without trend history
Viewing Results
After completing an evaluation, you'll see:
Overall Score: 78%
Status: B Grade
Category Breakdown:
• Usability: 82%
• Accessibility: 75%
• Performance: 73%
Detailed Results:
Question 1.1 [Usability] → Pass ✓
Question 1.2 [Usability] → Fail ✗
Question 2.1 [Accessibility] → Pass ✓
...This data feeds into your flow's Analytics page for comparison and trend analysis.
Best Practices
- Lock Guidelines Across Versions: Keep the same guideline set when continuing evaluations so results are comparable
- Add Notes: Include notes on each question to explain your assessment and document design decisions
- Regular Cadence: Run evaluations after design changes to establish a clear trend
- Use Analytics: After multiple evaluations, check Analytics to see progress patterns and identify which categories need attention