What This Template Is For
Building the wrong thing is the most expensive mistake a product team can make. Concept testing prevents it by putting ideas in front of users before you write a single line of code. You show participants a concept (a mockup, a storyboard, a description, or a rough prototype) and measure their reaction: Do they understand it? Do they want it? Would they pay for it? What is missing?
This template provides a structured concept testing protocol: stimulus preparation, evaluation criteria, discussion guide, reaction measurement scales, desirability scoring, and analysis framework. It works for testing features, product ideas, positioning, and pricing concepts.
The Product Discovery Handbook covers where concept testing fits in the validation process. If you are earlier in discovery and need to understand the problem space, start with the User Interview Script Template. For quantifying market demand after concept validation, the TAM Calculator helps size the opportunity. The minimum viable product glossary entry explains how concept testing informs your MVP scope.
How to Use This Template
- Prepare the stimulus. Create 1-3 concept representations. These can be mockups, wireframes, storyboards, product descriptions, or landing page prototypes. Keep them rough enough that participants focus on the concept, not the visual design.
- Define evaluation criteria. What does success look like? Set thresholds for comprehension, desirability, and purchase intent before running the test.
- Recruit 8-12 participants. Enough for qualitative patterns. Screen for your target user profile.
- Run moderated sessions. Show the concept without explaining it. Measure first reactions. Then probe with structured questions.
- Score reactions. Use the scales below to quantify qualitative feedback.
- Analyze and decide. Compare results against your success thresholds. Iterate, pivot, or proceed.
The Template
Section 1: Study Overview
| Field | Details |
|---|---|
| Study Name | [Descriptive name for this concept test] |
| Researcher | [Name and role] |
| Concept(s) Being Tested | [Brief description of 1-3 concepts] |
| Stimulus Type | Mockup / Wireframe / Storyboard / Description / Prototype |
| Number of Participants | [8-12 recommended] |
| Session Duration | [20-30 minutes per participant] |
| Format | Remote / In-person |
| Status | Planning / Recruiting / Testing / Analysis / Complete |
Section 2: Research Objectives
- ☐ Define what you are trying to learn about this concept
- ☐ Set success criteria before running the test (not after)
- ☐ Identify the go/no-go decision this test will inform
Objectives:
- [Do users understand what this concept does without explanation?]
- [Do users want this enough to use it / pay for it?]
- [What is missing, confusing, or concerning about this concept?]
Success Criteria (define before testing):
| Metric | Threshold for "Proceed" | Threshold for "Iterate" | Threshold for "Kill" |
|---|---|---|---|
| Comprehension | 80%+ understand without prompting | 50-79% understand | < 50% understand |
| Desirability | 70%+ rate as "very" or "extremely" desirable | 40-69% | < 40% |
| Purchase Intent | 60%+ would "definitely" or "probably" use/buy | 30-59% | < 30% |
Section 3: Stimulus Preparation
- ☐ Create concept representation(s) at appropriate fidelity
- ☐ Remove branding and polish that could bias reactions toward "liking the design"
- ☐ Test the stimulus with 1-2 colleagues to verify it communicates the concept
- ☐ Prepare a neutral introduction that does not sell the concept
Stimulus Checklist:
| Concept | Format | Fidelity | Key Elements Shown | Ready |
|---|---|---|---|---|
| [Concept A] | [Mockup / Description / Prototype] | [Low / Medium / High] | [What the participant will see] | [Y/N] |
| [Concept B] | [Mockup / Description / Prototype] | [Low / Medium / High] | [What the participant will see] | [Y/N] |
Fidelity guidelines:
- Low fidelity (sketches, descriptions): Best for early-stage ideas. Participants focus on the concept, not the execution.
- Medium fidelity (wireframes, clickable mockups): Good for testing interaction patterns and workflows.
- High fidelity (polished prototypes): Use only when visual design is part of the value proposition.
Section 4: Discussion Guide
| Phase | Time | Activity |
|---|---|---|
| Warm-up | 3 min | Background questions about current workflow and pain points |
| First exposure | 2 min | Show concept without explanation. Ask: "What do you think this is?" |
| Comprehension | 5 min | "In your own words, what does this do?" "Who is this for?" |
| Reaction | 5 min | "What is your first reaction?" "What stands out?" "What concerns you?" |
| Desirability scoring | 3 min | Administer desirability scale and purchase intent questions |
| Deep dive | 7 min | Probe specific elements: "Tell me about this section." "What would you expect to happen if you clicked here?" |
| Comparison (if testing multiple concepts) | 5 min | "Which concept do you prefer? Why?" |
| Wrap-up | 3 min | "What would need to be true for you to use this?" "What is missing?" |
Section 5: Measurement Scales
Comprehension (record after first exposure, before any explanation):
- ☐ 5 = Fully understood the concept and its value without prompting
- ☐ 4 = Understood the core idea with minor confusion on details
- ☐ 3 = Understood part of the concept but missed key elements
- ☐ 2 = Significant misunderstanding of the concept
- ☐ 1 = Did not understand what the concept was
Desirability (ask participant directly):
"How desirable is this concept to you?"
- ☐ 5 = Extremely desirable. I want this now
- ☐ 4 = Very desirable. I would use this regularly
- ☐ 3 = Somewhat desirable. Nice to have but not essential
- ☐ 2 = Slightly desirable. Might try it once
- ☐ 1 = Not at all desirable. I would not use this
Purchase Intent (ask participant directly):
"If this were available today, how likely would you be to [use it / upgrade / pay for it]?"
- ☐ 5 = Definitely would
- ☐ 4 = Probably would
- ☐ 3 = Might or might not
- ☐ 2 = Probably would not
- ☐ 1 = Definitely would not
Section 6: Participant Reaction Notes
| Participant | Comprehension (1-5) | Desirability (1-5) | Purchase Intent (1-5) | Key Quote | Concerns / Suggestions |
|---|---|---|---|---|---|
| P1 | [Score] | [Score] | [Score] | "[Quote]" | [Notes] |
| P2 | [Score] | [Score] | [Score] | "[Quote]" | [Notes] |
| ... | ... | ... | ... | ... | ... |
Section 7: Analysis and Decision
- ☐ Calculate average scores for comprehension, desirability, and purchase intent
- ☐ Compare averages against success criteria defined in Section 2
- ☐ Identify the most common concerns and suggestions
- ☐ Categorize feedback into: must-fix, nice-to-fix, and out-of-scope
- ☐ Make a go/iterate/kill decision based on the data
Results Summary:
| Metric | Average Score | Success Threshold | Result |
|---|---|---|---|
| Comprehension | [Avg] | [Threshold] | [Pass / Iterate / Fail] |
| Desirability | [Avg] | [Threshold] | [Pass / Iterate / Fail] |
| Purchase Intent | [Avg] | [Threshold] | [Pass / Iterate / Fail] |
Decision: [Proceed to build / Iterate on concept / Kill the concept]
Rationale: [2-3 sentences explaining the decision based on the data]
Filled Example: Testing a Collaborative Whiteboard Feature
Study Overview
| Field | Details |
|---|---|
| Study Name | Collaborative Whiteboard Concept Test |
| Researcher | Priya Sharma, Product Manager |
| Concept | A real-time collaborative whiteboard embedded in project workspaces |
| Stimulus | Medium-fidelity Figma prototype (5 screens) |
| Participants | 10 (PMs, designers, and engineering managers) |
| Session Duration | 25 minutes each |
Success Criteria (Set Before Testing)
| Metric | Proceed | Iterate | Kill |
|---|---|---|---|
| Comprehension | 80%+ score 4-5 | 50-79% | < 50% |
| Desirability | 70%+ score 4-5 | 40-69% | < 40% |
| Purchase Intent | 50%+ score 4-5 | 25-49% | < 25% |
Results
| Metric | Average | % Scoring 4-5 | Threshold | Result |
|---|---|---|---|---|
| Comprehension | 4.2 | 80% | 80% | Pass |
| Desirability | 3.8 | 60% | 70% | Iterate |
| Purchase Intent | 3.5 | 50% | 50% | Pass (borderline) |
Key Findings
| Theme | Frequency | Impact |
|---|---|---|
| Value clear for brainstorming sessions | 8/10 | Positive |
| Confused about whiteboard vs. existing document editor | 6/10 | Negative |
| Want templates (retrospective, user story map, flowchart) | 7/10 | Suggestion |
| Concerned about performance with large boards | 4/10 | Risk |
Decision. Iterate. Desirability fell short because participants did not see how the whiteboard was different from the existing document editor. Next steps: redesign the concept to show whiteboard templates and emphasize real-time cursor presence. Retest with 8 participants in two weeks.
Key Takeaways
- Set success criteria (comprehension, desirability, purchase intent thresholds) before testing, not after
- Show the concept without explaining it. First-impression comprehension is your most honest data point
- Use appropriate fidelity. Low fidelity for idea validation, medium for interaction patterns
- Score reactions quantitatively using the 5-point scales. Qualitative quotes provide context but scores drive decisions
- "Iterate" is a valid outcome. Most concepts need one or two rounds of refinement before proceeding
- Follow concept testing with usability testing once you decide to build
About This Template
Created by: Tim Adair
Last Updated: 3/4/2026
Version: 1.0.0
License: Free for personal and commercial use
