design QA
How to conduct a heuristic evaluation with a design QA tool
A step-by-step guide to running structured heuristic evaluations using a browser-based design QA tool.
Heurio Team
May 11, 202613 min read

Most teams know they should evaluate their interfaces before shipping. Few actually do it in a structured way. A heuristic evaluation gives you a repeatable method for catching usability problems, but the method only works when paired with the right design QA tool and a clear process.
A heuristic evaluation is a structured inspection method where evaluators judge a user interface against a set of recognized usability principles (heuristics) to identify design problems before real users encounter them.
We built Heurio because we kept watching teams try to do heuristic evaluations with spreadsheets, screenshots, and Slack threads. The findings got lost. Context disappeared. Developers couldn't reproduce the issues. This guide walks through how to run a heuristic evaluation step by step, using browser-based tools that keep every finding attached to the exact element where the problem lives.
A heuristic evaluation works best with 3 to 5 independent evaluators reviewing the same interface against a chosen set of principles.
Using a browser-based design QA tool captures the full context (screenshot, URL, device data, console logs) that spreadsheets miss.
Pick your heuristic framework before you start, not during the review.
Severity ratings turn a list of observations into a prioritized action plan developers can actually use.
Running evaluations on live or staging pages (not static mockups) catches implementation-specific problems that Figma reviews miss entirely.
- Heuristic evaluation
- A usability inspection method where experts review an interface against established design principles to find problems without user testing.
- Design QA tool
- Software that helps teams inspect, annotate, and report interface issues, often directly in the browser on live or staging pages.
- Severity rating
- A score assigned to each finding that combines the problem's frequency, impact, and persistence to guide prioritization.
- Browser-based bug reporting
- The practice of capturing and filing interface issues directly from a web browser, automatically including technical context like console logs and network data.
- Visual feedback tool
- A platform that lets reviewers attach comments, annotations, and screenshots to specific elements on a live webpage.
Why heuristic evaluations still matter for design QA
Jakob Nielsen and Rolf Molich introduced heuristic evaluation in 1990. Over three decades later, the method still works. Why? Because the core idea is simple: have knowledgeable people look at an interface and check it against known principles. You don't need a lab. You don't need participants. You need a framework and a process.
Nielsen Norman Group's research found that a single evaluator catches about 35% of usability issues in an interface. Add a second evaluator and coverage jumps. With five evaluators, you typically catch around 75% of problems. The method scales predictably.
But the method was designed in an era of desktop software and paper forms. Running it on modern web products requires modern tooling. Annotating a screenshot in PowerPoint doesn't cut it when you need to show a developer exactly which CSS class is rendering the wrong padding on a specific breakpoint.
Static reviews miss implementation bugs
Many teams still run "design reviews" inside Figma. They compare the design file to the built page side by side. This catches visual drift, sure. But it misses interaction states, loading behaviors, responsive breakpoints, and accessibility issues that only appear in a real browser.
A proper heuristic evaluation happens on the actual product. That means live pages or staging URLs. That's where a design QA tool that works in the browser becomes essential.
Choose your heuristic framework first
Before opening any tool, decide which set of heuristics you'll evaluate against. This is the step most teams skip. They jump straight into reviewing and end up with vague notes like "this feels off" or "the spacing is weird." A framework gives you specific criteria to judge against.
The most common choice is Nielsen's 10 Usability Heuristics. These cover visibility of system status, match between system and real world, user control and freedom, consistency, error prevention, recognition over recall, flexibility, aesthetic design, error recovery, and help documentation.
But Nielsen's list isn't the only option. Consider these alternatives depending on your focus:
Shneiderman's 8 Golden Rules emphasize consistency and informative feedback, useful for data-heavy applications.
WCAG 2.2 POUR Principles focus specifically on accessibility: perceivable, operable, understandable, and robust.
Bastien and Scapin's Ergonomic Criteria break down into more granular sub-criteria, useful for detailed evaluations.
Heurio recommends picking one framework per evaluation session. Mixing frameworks mid-review creates inconsistent findings and makes it harder to aggregate results across evaluators. We've found in our own QA workflow that teams get cleaner data when everyone evaluates against the same checklist.
Assemble your evaluation team
You need 3 to 5 evaluators. Fewer than three leaves too many gaps. More than five hits diminishing returns, as Nielsen Norman Group's analysis has consistently shown.
Who should evaluate? Ideally, a mix of roles. A designer catches visual inconsistencies and hierarchy problems. A developer spots interaction bugs and performance issues. A product manager or marketer notices confusing copy and unclear calls to action.
Each evaluator works independently. This is critical. If evaluators review together, the first person to speak biases everyone else. Independent evaluation, followed by a group merge session, produces more diverse findings.
Brief your evaluators on the framework
Don't assume everyone knows the heuristics by heart. Before the evaluation starts, share the chosen framework. Walk through each principle with one or two examples relevant to your product. This takes 15 minutes and dramatically improves finding quality.
If you're using Heurio, you can link evaluators directly to the heuristic evaluation guidelines page, which lists multiple frameworks with descriptions and examples. Each finding can be tagged to a specific heuristic, so the data stays structured.
Set up your design QA tool for the evaluation
This is where tooling makes or breaks the process. A design QA tool should capture context automatically. When an evaluator clicks on a problematic element, the tool should grab:
A screenshot of the current viewport
The URL and scroll position
Device and browser information
Console errors and network requests
The DOM selector of the element
This level of browser-based bug reporting context means developers can reproduce the issue without a back-and-forth Slack thread. Traditional visual feedback tools capture screenshots but often miss the technical layer. If you're looking for a better alternative that includes console logs alongside visual annotations, that's exactly the gap we built Heurio to fill.
Set up a project in your tool before the evaluation starts. Define the pages or flows to review. If you're evaluating a checkout flow, list the specific URLs: product page, cart, shipping, payment, confirmation. Give evaluators a clear scope.
Capture UX issues without leaving the browser
Heurio attaches contextual notes, screenshots, and console logs to any element on any page. Designers, developers, and vibe coders all use the same workflow.
Install the Heurio Chrome extension
Run the evaluation: a step-by-step design QA tool workflow
Here's the actual process we recommend. It works whether you're evaluating a SaaS dashboard, a marketing site, or a page generated by an AI design tool like Lovable or v0.
First pass: free exploration. Spend 10 to 15 minutes navigating the interface without taking notes. Get a sense of the overall flow, structure, and feel. Note your first impressions mentally but don't document yet.
Second pass: heuristic-by-heuristic review. Go through each heuristic in your chosen framework. For every principle, scan the interface looking specifically for violations. Click on the element in your design QA tool and log the finding with the heuristic tag, a description of the problem, and why it violates the principle.
Third pass: edge cases and states. Test error states, empty states, loading states, and responsive breakpoints. Resize the browser. Submit forms with invalid data. Click buttons rapidly. These interaction-layer issues are invisible in static mockups and often violate heuristics around error prevention and recovery.
Rate severity for each finding. Use a 0 to 4 scale: 0 means not a usability problem, 1 is cosmetic only, 2 is minor, 3 is major, and 4 is a usability catastrophe. Rate after documenting, not during. This prevents you from self-censoring minor issues.
Submit your independent findings. Each evaluator submits their annotated findings through the tool. No group discussion yet. Keep evaluations independent until everyone has submitted.
Merge and prioritize in a group session. Bring all evaluators together. Review overlapping findings (issues caught by multiple evaluators are almost always real problems). Average the severity ratings. Create a prioritized list sorted by severity.
What good findings look like
A bad finding: "The form is confusing."
A good finding: "The email input field on /signup does not show an error message when the user submits an invalid format. This violates Nielsen's heuristic #9 (help users recognize, diagnose, and recover from errors). Severity: 3. Screenshot and console log attached."
The difference is specificity. Good findings reference the exact element, the exact heuristic, and the exact severity. A design QA tool that forces this structure produces better data than a freeform spreadsheet.
Lovable visual feedback and why vibe coding makes evaluations more urgent
AI design tools ship interfaces fast. Vibe coding with tools like Lovable, Bolt, and Replit means you can go from prompt to deployed page in minutes. But speed creates a new problem: the interface looks polished enough to ship, yet it hasn't been evaluated against any usability principles.
A solo founder uses Lovable to generate a landing page and the page actually looks great. But the tap targets are too small for mobile, the error states are generic, the color contrast fails WCAG 2.2 AA requirements. These are exactly the issues a heuristic evaluation catches in 30 minutes.
Lovable visual feedback works best when you run it through a structured process. Don't just eyeball the page. Open your design QA tool, pick a heuristic framework, and do the three-pass review. The output is a list of specific, actionable issues you can feed right back into the AI tool or fix manually.
Bug reports with console logs close the loop faster
When your evaluation findings include console logs and network data, fixing becomes straightforward. A developer (or an AI coding assistant) can see the exact error, the exact state, and the exact element. This is what we mean by bug reports with console logs. The report isn't just "something broke." It's a complete reproduction package.
Compare that to the typical workflow: reviewer takes a screenshot, posts it in Slack, developer asks "what browser?", reviewer doesn't remember, developer asks "can you try again?", and two days pass before anyone looks at the actual problem.
Turning findings into a prioritized action plan
A heuristic evaluation produces a list of findings. That list is useless if it doesn't convert into tickets that developers actually fix. Here's how to bridge the gap.
First, sort by severity. All severity-4 issues (usability catastrophes) go into the current sprint. Severity-3 issues get scheduled for the next sprint. Severity-2 issues go into the backlog. Severity-1 and severity-0 issues get documented but don't block anything.
Second, group findings by heuristic. If you see five findings all violating "consistency and standards," that points to a systemic problem, not five isolated bugs. The fix might be a design system update rather than five individual patches.
Third, export findings to your project management tool. If your team uses Linear, Jira, or Notion, your visual feedback tool should integrate or export cleanly. Each finding becomes a ticket with the screenshot, severity, heuristic reference, and technical context already attached.
Google's Lighthouse audit can supplement your heuristic evaluation with automated accessibility and performance checks. Run Lighthouse on the same pages you evaluated manually. The automated scan catches the mechanical issues (missing alt text, contrast ratios, tap target sizes) while your heuristic evaluation catches the judgment-dependent problems (confusing flows, misleading affordances, poor error messaging).
Common mistakes when running heuristic evaluations
After working with dozens of teams on their QA processes, we see the same mistakes come up again and again.
Evaluating as a group instead of individually
Group evaluations feel efficient, but they're not. Research from NNGroup is clear: individual evaluations followed by a merge session produce more findings than group walkthroughs. The anchoring effect is real. If the most senior person says "this looks fine," junior evaluators stay quiet.
Skipping severity ratings
Without severity ratings, every finding looks equally important. A misaligned icon and a broken checkout button sit side by side in the report. Developers (correctly) lose trust in the evaluation because it doesn't distinguish between critical and cosmetic.
Evaluating mockups instead of live pages
Figma mockups don't have loading states, don't run JavaScript, don't render differently across browsers, and don't show real API errors. If you're doing design QA, do it in the browser. This is a hill we will absolutely stand on.
Using no framework at all
"Just look at the page and tell me what's wrong" is not a heuristic evaluation. It's an opinion session. The framework is what makes the method reliable and repeatable. Pick one. Stick with it. Your findings will be measurably better.
Frequently asked questions
How many evaluators do I need for a heuristic evaluation?
Three to five evaluators is the sweet spot. A single evaluator catches roughly 35% of issues. Five evaluators typically find around 75%. Adding more than five gives diminishing returns, so the cost-benefit ratio drops off sharply after that point.
What is the best design QA tool for heuristic evaluations?
The best design QA tool for heuristic evaluations works directly in the browser and captures full context: screenshots, console logs, device data, and DOM selectors. Heurio is purpose-built for this. Other options handle visual annotations but often lack the heuristic tagging and technical depth that structured evaluations require.
Can I use a design QA tool for accessibility evaluations too?
Yes. Accessibility evaluations overlap significantly with heuristic reviews, especially around error prevention, visibility of system status, and consistency. Combine a browser-based visual feedback tool with automated scanners like Lighthouse or axe DevTools for the most complete coverage.
How long does a heuristic evaluation take?
A single evaluator typically spends 60 to 90 minutes on a focused evaluation of one flow (5 to 10 screens). The group merge session adds another 30 to 60 minutes. Total wall-clock time for a five-person team evaluating one flow is usually one day, accounting for scheduling.
Should I run heuristic evaluations on AI-generated pages from Lovable or v0?
Absolutely. AI-generated pages often look polished but miss interaction details, accessibility requirements, and edge-case handling. A structured heuristic evaluation catches these gaps fast. Run the evaluation on the deployed or preview URL, not on the AI tool's editor view.
How often should teams run heuristic evaluations?
We recommend running a focused evaluation before every major release and a broader evaluation quarterly. Teams using continuous deployment benefit from lightweight heuristic checks on each significant feature branch. The key is making the practice habitual, not heroic.
Keep reading

Stop reviewing copy in Google Docs. Use a design QA tool
Copy review in Google Docs misses layout, truncation, and context. A design QA tool pins feedback to the real page.

Vibe coding workflow meets Nielsen's 10 usability heuristics
Use Nielsen's 10 usability heuristics as your QA checklist for AI-generated pages. A practical guide for vibe coders.

Design QA in the browser using Dieter Rams's 10 principles
Apply Dieter Rams's 10 design principles as a structured checklist for design QA in the browser.

