Back to skills
SkillHub ClubAnalyze Data & AIFull StackData / AI

product-analyst

Track user metrics and provide data-driven insights for product decisions. Use when measuring product health, analyzing user behavior, conducting cohort analysis, or optimizing key metrics. Covers acquisition, engagement, retention, revenue metrics, and data-driven decision making.

Packaged view

This page reorganizes the original catalog entry around fit, installability, and workflow context first. The original raw source lives below.

Stars
21
Hot score
87
Updated
March 20, 2026
Overall rating
C1.8
Composite score
1.8
Best-practice grade
B84.0

Install command

npx @skill-hub/cli install daffy0208-ai-dev-standards-product-analyst
product-analyticsmetricsdata-drivenuser-behaviorretention

Repository

daffy0208/ai-dev-standards

Skill path: skills/product-analyst

Track user metrics and provide data-driven insights for product decisions. Use when measuring product health, analyzing user behavior, conducting cohort analysis, or optimizing key metrics. Covers acquisition, engagement, retention, revenue metrics, and data-driven decision making.

Open repository

Best for

Primary workflow: Analyze Data & AI.

Technical facets: Full Stack, Data / AI.

Target audience: everyone.

License: Unknown.

Original source

Catalog source: SkillHub Club.

Repository owner: daffy0208.

This is still a mirrored public skill entry. Review the repository before installing into production workflows.

What it helps with

  • Install product-analyst into Claude Code, Codex CLI, Gemini CLI, or OpenCode workflows
  • Review https://github.com/daffy0208/ai-dev-standards before adding product-analyst to shared team environments
  • Use product-analyst for business workflows

Works across

Claude CodeCodex CLIGemini CLIOpenCode

Favorites: 0.

Sub-skills: 0.

Aggregator: No.

Original source / Raw SKILL.md

---
name: product-analyst
description: Track user metrics and provide data-driven insights for product decisions. Use when measuring product health, analyzing user behavior, conducting cohort analysis, or optimizing key metrics. Covers acquisition, engagement, retention, revenue metrics, and data-driven decision making.
---

# Product Analyst

Measure user behavior and product health to inform data-driven decisions.

## Core Principle

**What gets measured gets improved.** Define the right metrics, track them relentlessly, and act on insights quickly.

## North Star Metric

**The ONE metric that best captures value delivered to users.**

Your North Star should:

- ✅ Represent real customer value
- ✅ Correlate with revenue
- ✅ Be measurable frequently (daily/weekly)
- ✅ Rally the entire team around one goal

**Examples by Product Type**:

```yaml
Communication:
  Slack: Messages Sent (weekly active)
  Zoom: Weekly Meeting Minutes
  Discord: Active Servers

Marketplace:
  Airbnb: Nights Booked
  Uber: Completed Rides
  Etsy: Gross Merchandise Value (GMV)

Media/Content:
  Spotify: Time Listening
  Netflix: Hours Watched
  Medium: Total Time Reading

SaaS/B2B:
  Asana: Weekly Active Teams
  Notion: Collaborative Documents
  Salesforce: Deals Closed (CRM value)

Social:
  Facebook: Daily Active Users (DAU)
  Instagram: Posts Shared
  Twitter: Tweets per User
```

**How to choose your North Star**:

1. What action represents core value?
2. If users do this more, do they get more value?
3. Does this predict revenue?
4. Can the entire team influence it?

## Key Metrics by Category

### Acquisition Metrics

**Goal**: Get users into the product

```yaml
Traffic Sources:
  - Organic Search: SEO traffic
  - Paid Ads: Google Ads, Facebook Ads
  - Referral: Word of mouth, links
  - Direct: Typed URL, bookmarked
  - Social: Twitter, LinkedIn posts

Key Metrics:
  - Unique Visitors: Total website visitors
  - Sign-ups: Users who created account
  - Conversion Rate: Visitors → Sign-ups
  - Cost Per Acquisition (CPA): Ad spend / sign-ups
  - Source Quality: Which sources convert best?

Targets:
  - Visitor → Sign-up: 2-5% (good), 5-10% (excellent)
  - CPA: < $50 (B2C), < $200 (B2B), depends on LTV
```

### Activation Metrics

**Goal**: Get users to "aha moment"

```yaml
Activation Definition:
  - User completes onboarding
  - User takes first core action
  - User experiences product value

Examples:
  Slack: Sent 2,000 messages (team is active)
  Dropbox: Added file to folder
  Twitter: Followed 30 accounts
  Airbnb: Completed first booking

Key Metrics:
  - Activation Rate: Sign-ups → Activated
  - Time to Activation: How long to aha moment?
  - Onboarding Completion: % who finish setup

Targets:
  - Activation Rate: >40% (good), >60% (excellent)
  - Time to Activation: <24 hours (ideal)
```

### Engagement Metrics

**Goal**: Keep users coming back

```yaml
Key Metrics:
  - Daily Active Users (DAU)
  - Weekly Active Users (WAU)
  - Monthly Active Users (MAU)
  - DAU/MAU Ratio (Stickiness): How often users return
  - Session Frequency: Times per week user logs in
  - Session Duration: Time spent per visit
  - Feature Adoption: % using each feature

DAU/MAU Stickiness:
  Excellent: >40% (Facebook, Slack)
  Good: 20-40% (most SaaS)
  Needs Work: <20%

Session Frequency Targets:
  B2C Social: 5-7 times per week
  B2B Tools: 3-5 times per week
  E-commerce: 1-2 times per week
```

### Retention Metrics

**Goal**: Prevent churn

```yaml
Cohort Retention:
  - Day 1: % still active 1 day after sign-up
  - Day 7: % still active 7 days after
  - Day 30: % still active 30 days after

Good Retention Curves:
  Consumer B2C:
    - D1: 60-80%
    - D7: 40-60%
    - D30: 30-50%
    - Flattening curve (good!)

  Enterprise B2B:
    - D1: 80-90%
    - D7: 70-80%
    - D30: 60-70%
    - Very flat curve

Bad Retention:
  - D1: 40%
  - D7: 10%
  - D30: 2%
  - Steep drop-off = product-market fit issue

Churn Rate:
  - Monthly Churn: % users who stop using each month
  - Target: <5% (consumer), <1% (enterprise)
  - Churn = Revenue Leak

Net Retention:
  - (Starting Users + New - Churned) / Starting Users
  - Target: >100% (growth despite churn)
```

### Revenue Metrics

**Goal**: Monetize effectively

```yaml
Key Metrics:
  - MRR (Monthly Recurring Revenue): Predictable monthly income
  - ARR (Annual Recurring Revenue): MRR × 12
  - ARPU (Average Revenue Per User): Revenue / # users
  - LTV (Lifetime Value): Total revenue from user over lifetime
  - CAC (Customer Acquisition Cost): Sales + marketing / new customers
  - LTV:CAC Ratio: Must be > 3:1
  - Payback Period: Months to recover CAC

Calculations:
  LTV = ARPU × Average Lifetime (months)
  Average Lifetime = 1 / Churn Rate

  Example:
    ARPU: $50/month
    Churn: 5% per month
    Average Lifetime: 1 / 0.05 = 20 months
    LTV: $50 × 20 = $1,000

  CAC: $300
  LTV:CAC = $1,000 / $300 = 3.3:1 (Good!)

Targets:
  - LTV:CAC: >3:1 (minimum), >4:1 (healthy)
  - Payback Period: <12 months
  - MRR Growth: >10% month-over-month (early stage)
```

### Satisfaction Metrics

**Goal**: Keep customers happy

```yaml
NPS (Net Promoter Score):
  Question: "How likely are you to recommend us?" (0-10)
  - Promoters: 9-10
  - Passives: 7-8
  - Detractors: 0-6

  NPS = % Promoters - % Detractors

  Benchmarks:
    Excellent: >50
    Good: 30-50
    Needs Work: <30

CSAT (Customer Satisfaction):
  Question: "How satisfied are you?" (1-5)

  Target: >4.0 average

CES (Customer Effort Score):
  Question: "How easy was it to [task]?" (1-7)

  Target: <3.0 (low effort)
```

## Segmentation

**Don't treat all users the same.** Different cohorts behave differently.

```yaml
Segment by Engagement:
  Power Users (Top 10%):
    - Use daily
    - High engagement
    - Understand product deeply
    → Interview them for feature ideas

  Casual Users (Middle 60%):
    - Use occasionally
    - Basic feature adoption
    → What prevents them from power usage?

  At-Risk Users (Bottom 20%):
    - Haven't logged in 7+ days
    - Low engagement
    → Re-engagement campaign

  Churned Users:
    - No activity 30+ days
    → Exit survey, understand why

Segment by Acquisition Source:
  - Organic vs Paid
  - Which source has best retention?
  - Which source has best LTV?

Segment by Plan:
  - Free vs Paid
  - Starter vs Pro vs Enterprise
  - Which tier has best retention?

Segment by Cohort (Sign-up Date):
  - Week 1 users vs Week 2 users
  - Did product changes improve metrics?
```

## Funnel Analysis

**Track conversion at each stage:**

```yaml
Sign-up Funnel Example:
  1. Land on homepage:        10,000 users (100%)
  2. Click "Sign Up":          2,000 users (20%)
  3. Fill sign-up form:        1,200 users (12%)
  4. Verify email:               800 users (8%)
  5. Complete onboarding:        400 users (4%)

Analysis:
  Biggest drop-off: Homepage → Sign Up (80% lost)
  Fix: Clarify value prop, add social proof, improve CTA

  Second drop-off: Form → Email verify (33% lost)
  Fix: Simplify form, reduce friction

Optimize biggest drop-offs first for max impact.
```

## Cohort Analysis

**Compare user groups over time:**

```yaml
Example: Retention by Sign-up Week

Week 1 Cohort (Jan 1-7):
  100 users signed up
  - D1: 80 active (80%)
  - D7: 40 active (40%)
  - D30: 20 active (20%)

Week 2 Cohort (Jan 8-14):
  120 users signed up
  - D1: 102 active (85%)  ← +5% improvement!
  - D7: 60 active (50%)   ← +10% improvement!
  - D30: 36 active (30%)  ← +10% improvement!

Insight: Onboarding changes in Week 2 improved retention!

Action: Roll out Week 2 changes to all users.
```

## A/B Testing

**Test hypotheses systematically:**

```yaml
1. Form Hypothesis: 'Adding social proof to homepage will increase sign-ups by 10%'

2. Design Experiment:
  - Control: Current homepage
  - Treatment: Homepage + customer testimonials
  - Split: 50/50 traffic
  - Primary Metric: Sign-up rate
  - Duration: 2 weeks or 1,000 visitors per variant

3. Run Test:
  - Don't peek early (wait for significance)
  - Monitor for bugs/issues

4. Analyze Results:
  Control: 1,000 visitors → 20 sign-ups (2.0%)
  Treatment: 1,000 visitors → 25 sign-ups (2.5%)

  Lift: +25% relative
  P-value: 0.04 (significant at p<0.05)

  Decision: WIN - Ship it!

5. Document Learning: 'Social proof increases sign-ups by 25%. Apply to all high-intent pages.'

Minimum Sample Size:
  - 100+ conversions per variant minimum
  - More is better for small effects
```

## Dashboard Design

### Executive Dashboard

```yaml
Top Metrics (Big Numbers):
  - North Star Metric: 12,500 WAU
  - MRR: $42,000 (+12% MoM)
  - Users: 1,850 (+15% MoM)

Graphs (Trends):
  - North Star over time
  - Revenue growth
  - User acquisition

Alerts:
  - Churn spike: +20% this week ⚠️
  - Trial conversion down: 10% → 8% ⚠️
```

### Product Dashboard

```yaml
Engagement:
  - DAU: 3,200
  - WAU: 8,500
  - MAU: 15,000
  - Stickiness (DAU/MAU): 21%

Feature Usage:
  - Feature A: 80% adoption
  - Feature B: 45% adoption
  - Feature C: 12% adoption (low!)

Retention:
  - D1: 75%
  - D7: 50%
  - D30: 35%

Funnels:
  - Sign-up → Activation: 45%
  - Trial → Paid: 12%
```

### Marketing Dashboard

```yaml
Acquisition:
  - Visitors: 50,000
  - Sign-ups: 2,000 (4% conversion)
  - Activated: 800 (40% activation)

By Source:
  - Organic: 20,000 visitors, 5% conversion
  - Paid: 15,000 visitors, 3% conversion
  - Referral: 10,000 visitors, 6% conversion (best!)

Cost Efficiency:
  - CPA: $150
  - LTV: $600
  - LTV:CAC: 4:1 (healthy!)
```

## Tools & Software

```yaml
Event Tracking:
  - Mixpanel (best for product analytics)
  - Amplitude (great alternative)
  - PostHog (open-source)
  - Google Analytics 4 (free, basic)

Session Recording:
  - FullStory (see user sessions)
  - LogRocket (debugging + analytics)
  - Hotjar (heatmaps + recordings)

A/B Testing:
  - Optimizely
  - VWO
  - Google Optimize (free, basic)
  - LaunchDarkly (feature flags + testing)

Data Warehouse:
  - Snowflake
  - BigQuery
  - Redshift

Visualization:
  - Tableau
  - Looker
  - Metabase (open-source)
```

## Reporting Cadence

```yaml
Daily:
  - Check North Star Metric
  - Monitor error rates
  - Review yesterday's experiments

Weekly:
  - Funnel analysis
  - Cohort retention
  - Feature adoption
  - Share insights with team

Monthly:
  - MRR/ARR review
  - LTV:CAC ratio
  - Churn analysis
  - Send NPS survey

Quarterly:
  - Deep dive on user segments
  - Competitive benchmarking
  - Strategic planning with leadership
```

## Quick Start Checklist

- [ ] Define North Star Metric
- [ ] Set up event tracking (Mixpanel/Amplitude)
- [ ] Instrument key events (sign-up, activation, core actions)
- [ ] Create acquisition funnel
- [ ] Track retention cohorts
- [ ] Build executive dashboard
- [ ] Set up weekly reporting
- [ ] Run first A/B test

## Common Pitfalls

❌ **Vanity metrics**: Tracking metrics that look good but don't predict success (e.g., page views)
❌ **Too many metrics**: Focus on 3-5 key metrics, not 50
❌ **No North Star**: Team pulls in different directions
❌ **Ignoring segments**: Averages hide important patterns
❌ **Analysis paralysis**: Measure, learn, act quickly
❌ **Not acting on data**: Data without action is worthless

## Summary

Great product analysis:

- ✅ One North Star Metric everyone tracks
- ✅ AARRR framework (Acquisition, Activation, Retention, Revenue, Referral)
- ✅ Cohort analysis over time
- ✅ Segmentation (not all users are the same)
- ✅ Regular A/B testing
- ✅ Share insights widely with team
- ✅ Act on data quickly
product-analyst | SkillHub