Most product teams discover their analytics are broken only when a stakeholder asks a question they cannot answer. Events fire inconsistently, properties are missing, and nobody knows which dashboards to trust. This template gives you a structured process to audit every layer of your analytics stack, from event collection to dashboard accuracy.
Whether you are joining a new team or preparing for a major product launch, an analytics audit prevents costly decisions made on bad data. Teams that use the Product Analytics Handbook as a reference alongside this audit template typically surface 15-30 tracking issues in their first pass. If you are unfamiliar with the core concepts, start with the cohort analysis glossary entry to ground yourself in how analysts slice user behavior.
This template works for any analytics platform (Amplitude, Mixpanel, PostHog, Heap, GA4). The goal is not to check every event individually but to verify that your measurement architecture is sound and your most critical metrics are trustworthy.
When to Use This Template
Joining a new product team. Run this audit in your first two weeks to understand what data you can and cannot trust.
Before a major launch. Verify that all new feature events are instrumented correctly before you ship and lose pre-launch baseline data.
After a platform migration. Switching from one analytics tool to another introduces silent breakages that only an audit catches.
Quarterly data hygiene. Schedule a recurring audit each quarter to catch drift before it compounds.
When stakeholders distrust the data. If PMs or executives question your numbers, an audit gives you a credible baseline to rebuild confidence.
Post-incident review. After discovering a major tracking bug, audit the surrounding events to check for related issues.
How to Use This Template
Step 1: Inventory Your Events
Export your full event list from your analytics platform. For each event, note the last time it fired, the volume, and the properties attached. Flag any events with zero volume in the last 30 days.
Step 2: Map Events to Key Metrics
Identify your North Star metric and the 5-10 supporting metrics your team reviews weekly. Trace each metric back to the specific events that power it. Any metric without a clear event trail is a red flag.
Step 3: Validate Data Accuracy
Pick 3-5 high-stakes events and manually verify them. Trigger the event yourself, check that it appears in your analytics tool with the correct properties, and compare the volume against a server-side source of truth (database queries, backend logs).
Step 4: Check for Gaps and Redundancies
Look for user actions that matter but have no event. Look for duplicate events that measure the same thing with different names. Document both in the template below.
Step 5: Score and Prioritize Fixes
Rate each issue by severity (blocks a key metric vs. nice-to-have) and effort to fix. Use a simple RICE score if you need to negotiate engineering time for instrumentation work.
Audit your analytics quarterly. Tracking drift is inevitable as code changes accumulate.
Start by mapping your key metrics to specific events. If you cannot trace a metric to an event, it is not measurable.
Validate high-stakes events manually. Automated volume checks miss silent data corruption where events fire but properties are wrong.
Prioritize fixes that block key metric measurement over cosmetic cleanup.
Document every audit in a shared location so future team members understand what was checked and when.
Treat your event taxonomy as a product. It needs an owner, a review process, and a deprecation policy.
Frequently Asked Questions
How often should I run an analytics audit?+
Quarterly is the right cadence for most teams. If you ship frequently (daily deploys), consider a lightweight monthly check on your top 10 events and a full audit each quarter.
What is a good analytics health score?+
Aim for 80/100 or above. Below 70 means you likely have metrics that are unreliable. Below 50 means critical business decisions are probably being made on flawed data.
Should I audit every single event?+
No. Focus on events tied to your core metrics first. A typical SaaS product has 50-200 events, but only 15-30 directly power the metrics your team reviews weekly. Audit those thoroughly and spot-check the rest.
Who should own the analytics audit?+
The product analyst or data team lead is the natural owner. But the PM should co-own the prioritization of fixes, since they understand which metrics matter most for upcoming decisions.
How do I get engineering time to fix tracking issues?+
Frame fixes in terms of business impact. "We cannot measure activation rate" is more compelling than "event X is broken." Use severity ratings from the audit to negotiate priority alongside feature work.
Explore More Templates
Browse our full library of PM templates, or generate a custom version with AI.