Skip to main content

Agent Analytics

Skill Verified Active

Analytics your AI agent can actually use. Track, analyze, run A/B experiments, and optimize across all your projects via CLI. Includes a growth playbook so your agent knows HOW to grow, not just what to track.

Purpose

To empower AI agents with advanced analytics capabilities, enabling them to understand project performance, conduct experiments, and drive growth.

Features

  • Track project performance via CLI
  • Analyze traffic, conversions, and engagement
  • Run A/B experiments programmatically
  • Optimize projects based on analytics insights
  • Includes a growth playbook for agent guidance

Use Cases

  • Adding analytics tracking to new websites or apps
  • Monitoring project health and identifying growth opportunities
  • Testing different versions of headlines or calls-to-action
  • Performing funnel analysis and cohort retention studies

Non-Goals

  • Providing GUI dashboards
  • Managing user accounts or billing
  • Storing Personally Identifiable Information (PII)

Installation

First, add the marketplace

/plugin marketplace add davepoon/buildwithclaude
/plugin install all-skills@buildwithclaude

Quality Score

Verified
97 /100
Analyzed about 16 hours ago

Trust Signals

Last commit5 days ago
Stars2.9k
LicenseMIT
Status
View Source

Similar Extensions

Experimentation Platform Orchestrator

98

A platform decision framework for experimentation. When to use Statsig vs PostHog vs GrowthBook vs Optimizely vs Amplitude vs Eppo vs Kameleoon. How to migrate between them. How to coordinate when multi-platform is genuinely warranted. The decisions that compound for years and the ones you can defer. Triggers on which experimentation platform, choose Statsig vs PostHog, evaluate experimentation tools, switch experimentation platform, migrate from Optimizely, consolidate experimentation tools, multi-platform experimentation, experimentation platform decision, ab test platform selection, feature flag platform vs experiment platform, warehouse-native experiments, vendor lock-in experimentation. Also triggers when a team is asking about cost, governance, or migration cost across experimentation tools, or when an evaluation is starting.

Skill
rampstackco

Measure Experiment Design

100

Designs an A/B test or experiment with clear hypothesis, variants, success metrics, sample size, and duration. Use when planning experiments to validate product changes or test hypotheses.

Skill
product-on-purpose

Game Analytics Setup

100

Invoke when the user needs to set up analytics, define telemetry events, establish KPIs, build dashboards, configure A/B testing, or implement data-driven design capabilities. Triggers on: "analytics", "telemetry", "KPIs", "metrics", "player data", "retention", "DAU", "dashboard", "A/B testing", "funnel analysis". Do NOT invoke for balance tuning (use game-balance-check) or economy design (use game-economy-designer). Part of the AlterLab GameForge collection.

Skill
AlterLab-IEU

Customer Success Manager

100

Monitors customer health, predicts churn risk, and identifies expansion opportunities using weighted scoring models for SaaS customer success. Use when analyzing customer accounts, reviewing retention metrics, scoring at-risk customers, or when the user mentions churn, customer health scores, upsell opportunities, expansion revenue, retention analysis, or customer analytics. Runs three Python CLI tools to produce deterministic health scores, churn risk tiers, and prioritized expansion recommendations across Enterprise, Mid-Market, and SMB segments.

Skill
alirezarezvani

Referral Program

100

When the user wants to design, launch, or optimize a referral or affiliate program. Use when they mention 'referral program,' 'affiliate program,' 'word of mouth,' 'refer a friend,' 'incentive program,' 'customer referrals,' 'brand ambassador,' 'partner program,' 'referral link,' or 'growth through referrals.' Covers program mechanics, incentive design, and optimization — not just the idea of referrals but the actual system.

Skill
alirezarezvani

Experiment Design

99

A discipline for designing experiments (A/B tests, multivariate, holdouts) so the results actually answer the question you asked. Hypothesis writing, sample size, duration, segment analysis, interpretation, decision-making, and the common failure modes that produce confidently wrong shipping decisions.

Skill
rampstackco

© 2025 SkillRepo · Find the right skill, skip the noise.