跳转到主要内容
此内容尚未提供您的语言版本,正在以英文显示。

Measure Experiment Results

技能 已验证 活跃
属于:PM Skills

Documents the results of a completed experiment or A/B test with statistical analysis, learnings, and recommendations. Use after experiments conclude to communicate findings, inform decisions, and build organizational knowledge.

目的

To standardize and professionalize the documentation of experiment results, turning raw data into actionable organizational knowledge for better decision-making.

功能

  • Documents statistical analysis, learnings, and recommendations
  • Provides a structured output template
  • Includes a comprehensive example for guidance
  • Adheres to best practices for experiment reporting

使用场景

  • After an A/B test or experiment reaches statistical significance
  • To communicate findings to stakeholders
  • During decisions about feature shipping, iteration, or discontinuation
  • To build a repository of learnings for future experiments

非目标

  • Conducting the experiment itself
  • Collecting raw experiment data
  • Performing the statistical analysis (assumes analysis is complete)

工作流

  1. Summarize the Experiment
  2. Restate the Hypothesis
  3. Present Primary Results
  4. Analyze Secondary Metrics
  5. Segment the Data
  6. Extract Learnings
  7. Make a Recommendation
  8. Define Next Steps

Trust

  • info:Issues Attention14 issues opened and 11 closed in the last 90 days indicates active engagement, but the closure rate is below the pass threshold.

安装

请先添加 Marketplace

/plugin marketplace add product-on-purpose/pm-skills
/plugin install pm-skills@pm-skills-marketplace

质量评分

已验证
95 /100
1 day ago 分析

信任信号

最近提交2 days ago
星标205
许可证Apache-2.0
状态
查看源代码

类似扩展

Measure Experiment Design

100

Designs an A/B test or experiment with clear hypothesis, variants, success metrics, sample size, and duration. Use when planning experiments to validate product changes or test hypotheses.

技能
product-on-purpose

Experiment Design

99

A discipline for designing experiments (A/B tests, multivariate, holdouts) so the results actually answer the question you asked. Hypothesis writing, sample size, duration, segment analysis, interpretation, decision-making, and the common failure modes that produce confidently wrong shipping decisions.

技能
rampstackco

Brainstorm Experiments New

100

Design lean startup experiments (pretotypes) for a new product. Creates XYZ hypotheses and suggests low-effort validation methods like landing pages, explainer videos, and pre-orders. Use when validating a new product idea, creating pretotypes, or testing market demand.

技能
phuryn

Experimentation Platform Orchestrator

98

A platform decision framework for experimentation. When to use Statsig vs PostHog vs GrowthBook vs Optimizely vs Amplitude vs Eppo vs Kameleoon. How to migrate between them. How to coordinate when multi-platform is genuinely warranted. The decisions that compound for years and the ones you can defer. Triggers on which experimentation platform, choose Statsig vs PostHog, evaluate experimentation tools, switch experimentation platform, migrate from Optimizely, consolidate experimentation tools, multi-platform experimentation, experimentation platform decision, ab test platform selection, feature flag platform vs experiment platform, warehouse-native experiments, vendor lock-in experimentation. Also triggers when a team is asking about cost, governance, or migration cost across experimentation tools, or when an evaluation is starting.

技能
rampstackco

Ab Test Setup

98

When the user wants to plan, design, or implement an A/B test or experiment, or build a growth experimentation program. Also use when the user mentions "A/B test," "split test," "experiment," "test this change," "variant copy," "multivariate test," "hypothesis," "should I test this," "which version is better," "test two versions," "statistical significance," "how long should I run this test," "growth experiments," "experiment velocity," "experiment backlog," "ICE score," "experimentation program," or "experiment playbook." Use this whenever someone is comparing two approaches and wants to measure which performs better, or when they want to build a systematic experimentation practice. For tracking implementation, see analytics-tracking. For page-level conversion optimization, see page-cro.

技能
coreyhaines31

Data Warehouse Experimentation

97

Running experiments out of the data warehouse instead of via dedicated experiment platforms. SQL-based assignment, exposure logging discipline, metric definitions in dbt models, statistical analysis in SQL or Python, variance reduction with CUPED, sequential testing, and the operational tradeoffs vs platforms like Statsig and Optimizely. Triggers on warehouse-native experimentation, run experiments in BigQuery, run experiments in Snowflake, dbt experiments, SQL t-test, CUPED variance reduction, exposure log, sample ratio mismatch, sequential testing, mSPRT, doubly robust estimation, build vs buy experimentation. Also triggers when the team is choosing between platform and warehouse, building warehouse-native experiment infrastructure, auditing one, or running an experiment with a custom metric the platform cannot handle.

技能
rampstackco