>
product-analytics
Expert guidance for product analytics, helping product teams define metrics, build funnels, analyze retention, run A/B tests, and make data-driven decisions. Applies frameworks for North Star metrics, pirate metrics (AARRR), cohort analysis, and experiment design.
#analytics#metrics#retention#cohorts#funnels
terminal-skillsv1.0.0
Works with:claude-codeopenai-codexgemini-clicursor
Usage
$
✓ Installed product-analytics v1.0.0
Getting Started
- Install the skill using the command above
- Open your AI coding agent (Claude Code, Codex, Gemini CLI, or Cursor)
- Reference the skill in your prompt
- The AI will use the skill's capabilities automatically
Example Prompts
- "Generate a professional invoice for the consulting work done in January"
- "Draft an NDA for our upcoming partnership with Acme Corp"
Documentation
Overview
Product analytics, helping product teams define metrics, build funnels, analyze retention, run A/B tests, and make data-driven decisions. This skill applies frameworks for North Star metrics, pirate metrics (AARRR), cohort analysis, and experiment design.
Instructions
North Star Metric
markdown
## Define Your North Star Metric
The North Star is a single metric that captures the core value
your product delivers to customers. It aligns every team around
what matters most.
### Criteria for a Good North Star
1. Reflects customer value (not just revenue)
2. Leading indicator of revenue (not lagging)
3. Measurable and actionable
4. Every team can influence it
### Examples by Product Type
- **Marketplace**: Weekly transactions (Airbnb: nights booked)
- **SaaS Productivity**: Weekly active users completing core action
(Slack: messages sent in channels with 3+ participants)
- **Subscription Media**: Weekly engaged time (Spotify: listening hours)
- **E-commerce**: Weekly purchases from repeat customers
- **Developer Tool**: Weekly API calls in production (Stripe, Twilio)
### North Star Framework
Your North Star has 3-5 input metrics that drive it:
North Star: "Weekly active teams completing core workflow"
Input metrics:
1. **Breadth**: New teams activated this week
2. **Depth**: Core workflows completed per team per week
3. **Frequency**: Days active per week per team
4. **Efficiency**: Time to complete core workflow
5. **Quality**: Workflow completion rate (started vs finished)
Each team owns an input metric. Together they drive the North Star.
AARRR Pirate Metrics
markdown
## Pirate Metrics Funnel
### Acquisition — How do users find you?
Metrics: Website visitors, signup rate, CAC, channel attribution
Questions: Which channels bring the highest-quality users? What's the CAC by channel?
### Activation — Do users experience core value?
Metrics: Onboarding completion rate, time-to-first-value, "aha moment" reached
Questions: What % of signups complete onboarding? What's the "aha moment"?
### Retention — Do users come back?
Metrics: D1/D7/D30 retention, weekly active rate, churn rate
Questions: What does the retention curve look like? Where does it flatten?
### Revenue — Do users pay?
Metrics: Conversion rate (free → paid), ARPU, LTV, expansion revenue
Questions: What triggers the upgrade? What's the payback period on CAC?
### Referral — Do users invite others?
Metrics: Viral coefficient, referral rate, NPS, organic share rate
Questions: Do users invite others? What's the average referrals per user?
### Identify Your Bottleneck
Measure each stage. The stage with the worst drop-off is your bottleneck.
Focus there — don't optimize acquisition if nobody activates.
Retention Analysis
markdown
## Analyze Retention
### Retention Curve
Plot % of users who return on Day 1, Day 7, Day 14, Day 30.
- **Good**: Curve flattens (users who stay past day 7 stick around)
- **Bad**: Curve approaches zero (no habitual users)
### Cohort Analysis
Compare retention across weekly or monthly sign-up cohorts:
Cohort | Week 1 | Week 2 | Week 3 | Week 4 | Week 8
Jan 1-7 | 100% | 42% | 31% | 28% | 25%
Jan 8-14 | 100% | 45% | 35% | 32% | 29%
Jan 15-21 | 100% | 51% | 40% | 38% | —
Reading this: Jan 15-21 cohort retains better → what changed?
Check: product changes, marketing channel mix, seasonality.
### Retention by Segment
Break retention by:
- **Acquisition channel**: Do SEO users retain better than paid?
- **Plan tier**: Do Pro users retain better than Free?
- **Activation actions**: Did they complete onboarding? Use feature X?
- **Company size**: Do small teams churn more than large?
### Find the "Aha Moment"
The aha moment is the action that predicts retention.
Method:
1. List all actions a user can take in the first week
2. For each action, split users who did it vs didn't
3. Compare 30-day retention between the two groups
4. The action with the biggest retention gap is your aha moment
Example findings:
- Users who create 3+ projects in week 1: 72% D30 retention
- Users who create 0-2 projects: 18% D30 retention
→ Aha moment: creating the 3rd project
→ Action: optimize onboarding to drive users to create 3 projects
A/B Testing
markdown
## Run A/B Tests
### Before You Test
1. Define the hypothesis: "Changing X will improve Y by Z%"
2. Choose the primary metric (one!)
3. Calculate sample size: use a power calculator
- Baseline conversion: current rate
- Minimum detectable effect: smallest change worth detecting
- Statistical power: 80% (standard)
- Significance level: 95% (standard)
4. Estimate duration: sample size ÷ daily traffic
### Common Mistakes
- ❌ Stopping early because results "look significant"
- ❌ Testing too many variants (dilutes sample size)
- ❌ Changing the test mid-experiment
- ❌ Using the wrong metric (vanity vs actionable)
- ❌ Not segmenting results (overall flat, but +30% for mobile users)
### Interpreting Results
**Statistical significance ≠ practical significance**
- 2% lift with p<0.05 might not be worth the engineering cost
- Consider the confidence interval, not just the point estimate
- Always check for novelty effects (run for 2+ full weeks)
### Sequential Testing
For faster decisions, use sequential testing:
- Define stopping rules before starting
- Check daily, but only stop if the boundary is crossed
- Avoids the "peeking problem" of traditional tests
### Post-Test
- Document: hypothesis, variant, result, learnings
- If winner: roll out to 100%
- If flat: was the sample size large enough? Consider the learning.
- If loser: document why and share the learning
Funnel Analysis
markdown
## Build and Optimize Funnels
### Define the Funnel
Map every step from entry to conversion:
E-commerce: Visit → Product View → Add to Cart → Checkout → Purchase
SaaS: Visit → Signup → Onboarding → Activation → Upgrade
### Measure Drop-off
Step | Users | Conversion | Drop-off
Visit | 10,000 | 100% | —
Signup | 2,500 | 25% | 75%
Complete onboarding | 1,000 | 40% | 60%
Reach aha moment | 600 | 60% | 40%
Still active at D30 | 300 | 50% | 50%
Upgrade to paid | 90 | 30% | 70%
### Find the Biggest Lever
The step with the biggest absolute drop-off has the most room for improvement.
In the example above: 75% drop at Signup → fix the landing page first.
### Optimize Each Step
- **Visit → Signup**: Value proposition clarity, social proof, friction reduction
- **Signup → Onboarding**: Reduce form fields, add progress indicators
- **Onboarding → Activation**: Guide to aha moment, remove unnecessary steps
- **Activation → Retention**: Habit loops, notifications, value reminders
- **Retention → Revenue**: Upgrade triggers, usage limits, feature gating
Guidelines
- One North Star — Align the entire product team around a single metric that reflects customer value
- Input metrics per team — Each team owns an input metric that drives the North Star; this creates autonomy with alignment
- Retention before acquisition — Fix retention first; acquiring users into a leaky bucket wastes money
- Cohort everything — Never look at aggregate metrics; always break by cohort (signup week, plan, channel) to find patterns
- Find the aha moment — Identify the action that predicts retention; then optimize onboarding to drive users to that action
- A/B test with discipline — Pre-register hypothesis, sample size, and duration; never peek and stop early
- Instrument early — Add analytics from day one; retroactive instrumentation means lost data you can never recover
- Metrics are questions — A metric tells you WHAT happened; you need qualitative research (interviews, session recordings) to understand WHY
Information
- Version
- 1.0.0
- Author
- terminal-skills
- Category
- Business
- License
- Apache-2.0