What This Template Is For
Most product backlogs are lists of features. Features are solutions. The problem is that solution backlogs skip the most important question: "Is the underlying belief actually true?"
A hypothesis backlog flips the model. Instead of tracking features to build, you track beliefs to test. Each entry is a statement about users, their problems, or the impact of a proposed solution. Each hypothesis has a current evidence level, a priority score, and a next experiment. The backlog ensures your team is always working on the highest-risk, highest-value unknowns instead of building features based on untested assumptions.
This template provides the format for maintaining a living hypothesis backlog. It covers hypothesis writing (clear, testable statements), evidence tracking (what you already know and how confident you are), priority scoring (impact and uncertainty, similar to the assumption testing template), and experiment planning (the next cheapest test for each top hypothesis).
The hypothesis backlog is a core artifact of continuous discovery. The Product Discovery Handbook describes how to integrate it into your weekly rhythm alongside opportunity solution trees, interview programs, and experiment cycles.
When to Use This Template
- When adopting continuous discovery. The hypothesis backlog is the connective tissue between customer research and roadmap decisions. Start here.
- During quarterly planning. Review the backlog to identify which hypotheses have been validated (and can become roadmap items) and which are still untested (and need experiments before committing resources).
- After every customer interaction. Each customer interview, support ticket cluster, or analytics anomaly should generate or update hypotheses in the backlog.
- When the team is split on direction. Instead of debating opinions, add each perspective as a testable hypothesis and let evidence decide.
- When running multiple experiments simultaneously. The backlog provides a single view of all active experiments and their status.
How to Use This Template
- Seed the backlog. Extract 10-15 hypotheses from your current roadmap, recent customer interviews, support ticket themes, and team assumptions.
- Score and prioritize. Rate each hypothesis on impact and uncertainty. Sort by risk score.
- Assign experiments. For the top 5 hypotheses, define the next experiment.
- Review weekly. Spend 15-20 minutes per week updating evidence, reviewing experiment results, and promoting or retiring hypotheses.
- Feed the roadmap. Validated hypotheses become backlog items. Invalidated hypotheses are archived with their evidence.
The Template
Backlog Setup
| Field | Details |
|---|---|
| Product / Team | [Product name or team] |
| Backlog Owner | [PM or discovery lead] |
| Review Cadence | [Weekly recommended. Minimum: biweekly.] |
| Last Review Date | [YYYY-MM-DD] |
| Active Hypotheses | [Count] |
| In Experiment | [Count] |
| Validated This Quarter | [Count] |
| Invalidated This Quarter | [Count] |
Hypothesis Format
Every hypothesis in the backlog should follow this structure:
We believe that [specific belief about users, their problem, or a solution]
for [target user or segment]
which will result in [expected outcome]
We will know this is true when [measurable evidence criteria]
Types of hypotheses:
- Problem hypotheses. "We believe that [user segment] struggles with [problem] because [reason]."
- Solution hypotheses. "We believe that [feature/change] will [outcome] for [users] because [reasoning]."
- Growth hypotheses. "We believe that [acquisition/activation/retention change] will increase [metric] by [amount] because [reasoning]."
- Business model hypotheses. "We believe that [segment] will pay [price] for [value] because [evidence]."
The Backlog
| ID | Hypothesis (short) | Type | Target Segment | Impact (1-5) | Uncertainty (1-5) | Risk Score | Evidence Level | Status | Next Experiment |
|---|---|---|---|---|---|---|---|---|---|
| H-001 | Problem / Solution / Growth / Business | None / Weak / Moderate / Strong | Active / In Experiment / Validated / Invalidated / Parked | ||||||
| H-002 | |||||||||
| H-003 | |||||||||
| H-004 | |||||||||
| H-005 | |||||||||
| H-006 | |||||||||
| H-007 | |||||||||
| H-008 | |||||||||
| H-009 | |||||||||
| H-010 |
Scoring guide:
- Impact (1-5): If this hypothesis is true (or false), how much does it affect the product's success?
- 5 = Critical to the product's core value proposition
- 3 = Affects a significant feature or metric
- 1 = Minor nice-to-have
- Uncertainty (1-5): How little evidence do we currently have?
- 5 = Pure assumption, no supporting data
- 3 = Some anecdotal evidence (a few interviews, some data)
- 1 = Well-supported by multiple data sources
- Risk Score: Impact x Uncertainty. Higher = test sooner.
Evidence Register
For each hypothesis, maintain a running log of evidence collected.
H-[ID]: [Hypothesis short name]
Full hypothesis: [Complete statement in the We believe... format]
| Date | Evidence Source | Finding | Supports / Contradicts | Confidence |
|---|---|---|---|---|
| [YYYY-MM-DD] | [Interview / Survey / Analytics / Experiment / Support data] | [What you learned] | S / C | High / Med / Low |
Current evidence assessment:
- Supporting evidence: [Count and summary]
- Contradicting evidence: [Count and summary]
- Net confidence: [None / Weak / Moderate / Strong]
- Recommendation: [Test further / Validate / Invalidate]
Experiment Tracker
Track active and completed experiments linked to hypotheses.
| Exp ID | Hypothesis | Method | Start Date | End Date | Status | Result | Action Taken |
|---|---|---|---|---|---|---|---|
| E-001 | H-[ID] | [Interview / Survey / Prototype / Fake door / Data analysis] | Planned / Running / Complete | Pass / Fail / Inconclusive | [What happened next] | ||
| E-002 | |||||||
| E-003 |
Weekly Review Agenda
Use this checklist during your weekly hypothesis review (15-20 minutes):
- ☐ Review experiment results from the past week. Update evidence register.
- ☐ Update evidence levels for hypotheses with new data (interviews, support tickets, analytics).
- ☐ Re-score any hypotheses where evidence has shifted the uncertainty rating.
- ☐ Promote validated hypotheses to the feature backlog with evidence documentation.
- ☐ Archive invalidated hypotheses with their evidence (do not delete them).
- ☐ Assign experiments for the top 2-3 unaddressed high-risk hypotheses.
- ☐ Add new hypotheses from recent customer conversations or data discoveries.
Validated Hypothesis Archive
When a hypothesis is validated or invalidated, move it here with its evidence summary.
| ID | Hypothesis | Verdict | Key Evidence | Decision Made | Date |
|---|---|---|---|---|---|
| H-[ID] | Validated / Invalidated | [1-2 sentence summary] | [Feature added to roadmap / Idea killed / Direction changed] |
Filled Example: B2B Collaboration Tool
Context. A B2B document collaboration tool (Series A, 300 customers) maintains a hypothesis backlog to drive their discovery process. Here is a snapshot of their backlog after 6 weeks.
Backlog Snapshot (Example)
| ID | Hypothesis | Type | Impact | Uncertainty | Risk | Evidence | Status |
|---|---|---|---|---|---|---|---|
| H-001 | Users leave because they cannot find documents created by teammates | Problem | 5 | 3 | 15 | Moderate | In Experiment |
| H-002 | Real-time co-editing will reduce the email-attachment workflow by 80% | Solution | 4 | 4 | 16 | Weak | In Experiment |
| H-003 | Teams of 10+ need folder permissions, not just document permissions | Problem | 3 | 5 | 15 | None | Active |
| H-004 | A Slack integration will increase daily active usage by 20% | Growth | 3 | 4 | 12 | Weak | Active |
| H-005 | Enterprise buyers require SSO before purchasing | Business | 5 | 2 | 10 | Strong | Validated |
Evidence Register for H-001 (Example)
Full hypothesis: We believe that users leave because they cannot find documents created by teammates, which results in frustration and eventual churn.
| Date | Source | Finding | S/C | Confidence |
|---|---|---|---|---|
| 2026-01-15 | Support tickets | 23 tickets in Q4 mention "can't find" or "where is the doc" | S | High |
| 2026-01-22 | Analytics | 35% of search queries return 0 results. Avg search session: 3.2 attempts | S | High |
| 2026-02-05 | Interview (P1) | "I just Slack my teammate and ask for the link. Every time." | S | Med |
| 2026-02-05 | Interview (P2) | "Our team has a shared bookmark doc in Google Docs with links to all our [product] docs" | S | Med |
| 2026-02-12 | Churn survey | 4 of 12 churned accounts cited "hard to find things" as a factor | S | Med |
Assessment: 5 supporting signals, 0 contradicting. Net confidence: Moderate. Currently testing via tree test on proposed new navigation structure.
Key Takeaways
- A hypothesis backlog replaces "I think" with "We believe, and here is the evidence." It moves the team from opinion-driven to evidence-driven decision making.
- Seed the backlog by auditing your current roadmap. For each planned feature, ask: "What must be true for this to succeed?" Each answer is a hypothesis.
- Review weekly. A hypothesis backlog that is updated monthly is a graveyard. Weekly 15-minute reviews keep it alive and relevant.
- The goal is not to test every hypothesis. It is to test the ones with the highest risk score first. Low-impact or well-evidenced hypotheses can stay in the backlog without active experiments.
- Validated hypotheses become feature backlog items. The evidence log travels with them, giving engineers and designers context for why the feature matters. Use the RICE framework to prioritize validated hypotheses when converting them to roadmap items.
- Archive invalidated hypotheses with their evidence. They protect the team from revisiting dead ideas. Six months from now, when someone says "What about real-time co-editing?", you can point to the evidence log instead of re-debating.
- This template integrates well with customer discovery and feature validation workflows. Each interview generates hypotheses; each validation experiment resolves them.
About This Template
Created by: Tim Adair
Last Updated: 3/5/2026
Version: 1.0.0
License: Free for personal and commercial use
