What This Template Is For
Most competitive analyses stop at feature lists. "They have dark mode, we do not." That tells you nothing about whether their dark mode is any good, whether users care about it, or whether it creates a switching incentive. A competitive UX audit goes deeper: you evaluate the actual user experience of 3-5 competitors across specific workflows, scoring both feature presence and experience quality.
This template provides a structured framework for competitive UX auditing: competitor selection criteria, workflow-based evaluation matrix, experience scoring rubric, heuristic evaluation checklist, gap analysis, and opportunity identification. It is designed for PMs and designers who need to understand where competitors are strong, where they are weak, and where the market has unmet needs.
The Product Discovery Handbook covers how competitive analysis fits into the broader discovery process. For a structured approach to evaluating your own product's usability, see the usability testing glossary entry. If you need to compare PM software tools specifically, the PM Tool Picker provides a data-driven comparison across 40 products. For quantifying the market opportunity you find, the TAM Calculator helps size the gap.
How to Use This Template
- Select competitors to audit. Choose 3-5 direct and indirect competitors. Include one market leader, one fast-growing challenger, and one outside-category product that solves a similar job.
- Define workflows to evaluate. Pick 3-5 critical workflows (not features). "Creating a project and inviting teammates" is a workflow. "Has project creation" is a feature checkbox.
- Create test accounts. Sign up for each competitor using the same persona. Complete the same workflows in the same order.
- Score each workflow. Use the experience scoring rubric below. Score on both task completion and experience quality.
- Document with screenshots. Capture every step. Screenshots are evidence that makes your findings credible.
- Identify gaps and opportunities. Look for workflows where all competitors score poorly. Those are your biggest differentiation opportunities.
The Template
Section 1: Audit Overview
| Field | Details |
|---|---|
| Audit Name | [Descriptive name] |
| Auditor | [Name and role] |
| Date | [Date conducted] |
| Product Category | [e.g., Project management, CRM, Analytics] |
| Competitors Audited | [List 3-5 competitor names] |
| Workflows Evaluated | [List 3-5 workflows] |
| Persona Used | [Description of the test persona] |
| Status | Planning / In Progress / Analysis / Complete |
Section 2: Competitor Selection
- ☐ Include the market leader (sets user expectations)
- ☐ Include a fast-growing challenger (potential disruptor)
- ☐ Include an indirect competitor (different category, same job-to-be-done)
- ☐ Include your own product (honest self-assessment)
- ☐ Verify you can access each product (free trial, demo account, or paid plan)
| Competitor | Category | Pricing Tier | Market Position | Why Included |
|---|---|---|---|---|
| [Name] | [Direct / Indirect] | [Free / $X/mo] | [Leader / Challenger / Niche] | [Why audit this one?] |
| [Name] | [Direct / Indirect] | [Free / $X/mo] | [Leader / Challenger / Niche] | [Why audit this one?] |
| [Name] | [Direct / Indirect] | [Free / $X/mo] | [Leader / Challenger / Niche] | [Why audit this one?] |
| Your Product | Self | [Price] | [Position] | Baseline comparison |
Section 3: Workflow Evaluation Matrix
For each workflow, score every competitor on two dimensions: task completion (can the user do it?) and experience quality (how pleasant and efficient is it?).
Scoring rubric:
| Score | Task Completion | Experience Quality |
|---|---|---|
| 5 | Completes in minimal steps, no errors | Delightful. Anticipates needs, provides guidance |
| 4 | Completes with minor friction | Good. Clear, efficient, minimal confusion |
| 3 | Completes but requires workaround or help docs | Adequate. Gets the job done without delight |
| 2 | Partially completes. Missing steps or capabilities | Frustrating. Confusing labels, unclear paths |
| 1 | Cannot complete the workflow | Broken. Errors, dead ends, or unusable |
Workflow 1: [Name of workflow]
| Step | Your Product | Competitor A | Competitor B | Competitor C |
|---|---|---|---|---|
| [Step 1] | [Score] | [Score] | [Score] | [Score] |
| [Step 2] | [Score] | [Score] | [Score] | [Score] |
| [Step 3] | [Score] | [Score] | [Score] | [Score] |
| Total Steps | [Count] | [Count] | [Count] | [Count] |
| Time to Complete | [Minutes] | [Minutes] | [Minutes] | [Minutes] |
| Completion Score | [1-5] | [1-5] | [1-5] | [1-5] |
| Experience Score | [1-5] | [1-5] | [1-5] | [1-5] |
- ☐ Repeat this table for each workflow (3-5 workflows total)
- ☐ Capture screenshots at every step for each competitor
Section 4: Heuristic Evaluation
Score each competitor on Nielsen's 10 usability heuristics (1 = poor, 5 = excellent).
| Heuristic | Your Product | Competitor A | Competitor B | Competitor C |
|---|---|---|---|---|
| Visibility of system status | [1-5] | [1-5] | [1-5] | [1-5] |
| Match between system and real world | [1-5] | [1-5] | [1-5] | [1-5] |
| User control and freedom | [1-5] | [1-5] | [1-5] | [1-5] |
| Consistency and standards | [1-5] | [1-5] | [1-5] | [1-5] |
| Error prevention | [1-5] | [1-5] | [1-5] | [1-5] |
| Recognition over recall | [1-5] | [1-5] | [1-5] | [1-5] |
| Flexibility and efficiency | [1-5] | [1-5] | [1-5] | [1-5] |
| Aesthetic and minimalist design | [1-5] | [1-5] | [1-5] | [1-5] |
| Help users recognize and recover from errors | [1-5] | [1-5] | [1-5] | [1-5] |
| Help and documentation | [1-5] | [1-5] | [1-5] | [1-5] |
| Average | [Avg] | [Avg] | [Avg] | [Avg] |
Section 5: Feature Comparison Matrix
- ☐ List all features relevant to the workflows under evaluation
- ☐ Mark presence, absence, and quality for each competitor
| Feature | Your Product | Competitor A | Competitor B | Competitor C |
|---|---|---|---|---|
| [Feature 1] | [Yes/No + quality note] | [Yes/No + quality note] | [Yes/No + quality note] | [Yes/No + quality note] |
| [Feature 2] | [Yes/No + quality note] | [Yes/No + quality note] | [Yes/No + quality note] | [Yes/No + quality note] |
Section 6: Gap Analysis and Opportunities
- ☐ Identify workflows where ALL competitors score below 3 (market-wide gap)
- ☐ Identify workflows where competitors score 4-5 but your product scores below 3 (catch-up priority)
- ☐ Identify features competitors have but your product lacks (table stakes gaps)
- ☐ Identify features your product has that competitors lack (strengths to protect)
| Gap Type | Description | Opportunity | Priority |
|---|---|---|---|
| Market gap | [Workflow where all competitors are weak] | [Differentiation opportunity] | [High / Medium / Low] |
| Catch-up gap | [Where competitors are ahead of you] | [Feature/UX improvement needed] | [High / Medium / Low] |
| Table stakes gap | [Feature everyone else has, you do not] | [Must-build to remain competitive] | [High / Medium / Low] |
| Strength | [Where your product leads] | [Protect and amplify] | [Maintain] |
Section 7: Deliverables
- ☐ Executive summary (1 page: key findings, top 3 opportunities)
- ☐ Full scoring matrices with screenshots
- ☐ Gap analysis with prioritized recommendations
- ☐ Competitor screenshot library organized by workflow
- ☐ Recommendation deck for stakeholder review
Filled Example: Auditing Project Management Tool Onboarding
Competitors Audited
| Competitor | Category | Market Position |
|---|---|---|
| Asana | Direct | Leader |
| Linear | Direct | Challenger |
| Notion | Indirect | Horizontal tool with PM features |
| Our Product | Self | Mid-market challenger |
Workflow: First Project Setup + Team Invite
| Metric | Our Product | Asana | Linear | Notion |
|---|---|---|---|---|
| Steps to create first project | 7 | 4 | 3 | 6 |
| Time to complete | 4:30 | 2:15 | 1:45 | 3:30 |
| Completion Score | 3 | 5 | 5 | 4 |
| Experience Score | 2 | 4 | 5 | 3 |
Gap Analysis Summary
| Gap Type | Finding | Recommendation |
|---|---|---|
| Catch-up | Onboarding takes 2x longer than Linear | Reduce project creation to 3 steps, add smart defaults |
| Market gap | No competitor provides sample data in the first project | Pre-populate first project with example tasks and views |
| Table stakes | We lack project templates at signup | Add 5 starter templates (competitors offer 10-20) |
| Strength | Our collaboration features scored highest (4.5 avg) | Highlight in onboarding, show collaboration value early |
Key Takeaways
- Audit workflows, not feature lists. "Can the user complete the task?" matters more than "Does the feature exist?"
- Score both task completion and experience quality. A feature can exist but be poorly implemented
- Include one indirect competitor. Users compare you to everything they use, not just direct competitors
- Document with screenshots. Visual evidence makes findings credible and actionable
- Focus your roadmap on market-wide gaps (differentiation) and table stakes gaps (retention)
- Repeat lightweight audits quarterly to catch competitor releases
About This Template
Created by: Tim Adair
Last Updated: 3/4/2026
Version: 1.0.0
License: Free for personal and commercial use
