跳转到主要内容
此内容尚未提供您的语言版本,正在以英文显示。

Data Quality Frameworks

技能 已验证 活跃

Implement data quality validation with Great Expectations, dbt tests, and data contracts. Use when building data quality pipelines, implementing validation rules, or establishing data contracts.

目的

Implement comprehensive data quality validation and testing within data pipelines using industry-standard tools like Great Expectations, dbt, and data contracts.

功能

  • Implement data quality validation with Great Expectations
  • Build dbt test suites
  • Establish data contracts
  • Automate data validation pipelines
  • Monitor data quality metrics

使用场景

  • Building data quality pipelines
  • Implementing validation rules
  • Establishing data contracts between teams
  • Monitoring data quality metrics

非目标

  • Full data pipeline orchestration beyond quality checks
  • Development of the data quality tools themselves
  • Data transformation or ETL logic

工作流

  1. Understand data quality dimensions and testing pyramid.
  2. Set up Great Expectations (installation, datasource, expectations).
  3. Configure Great Expectations checkpoints for validation.
  4. Implement dbt tests for data models.
  5. Define and implement data contracts.
  6. Automate quality checks within a pipeline.

实践

  • Data Quality
  • Testing
  • Pipeline Development

先决条件

  • Python environment
  • Great Expectations installed
  • dbt installed

安装

请先添加 Marketplace

/plugin marketplace add wshobson/agents
/plugin install data-engineering@claude-code-workflows

质量评分

已验证
97 /100
1 day ago 分析

信任信号

最近提交3 days ago
星标35.3k
许可证MIT
状态
查看源代码

类似扩展

Data Quality Auditor

97

Audit datasets for completeness, consistency, accuracy, and validity. Profile data distributions, detect anomalies and outliers, surface structural issues, and produce an actionable remediation plan.

技能
alirezarezvani

Senior Data Engineer

95

Data engineering skill for building scalable data pipelines, ETL/ELT systems, and data infrastructure. Expertise in Python, SQL, Spark, Airflow, dbt, Kafka, and modern data stack. Includes data modeling, pipeline orchestration, data quality, and DataOps. Use when designing data architectures, building data pipelines, optimizing data workflows, implementing data governance, or troubleshooting data issues.

技能
alirezarezvani

Geofeed Tuner

100

Use this skill whenever the user mentions IP geolocation feeds, RFC 8805, geofeeds, or wants help creating, tuning, validating, or publishing a self-published IP geolocation feed in CSV format. Intended user audience is a network operator, ISP, mobile carrier, cloud provider, hosting company, IXP, or satellite provider asking about IP geolocation accuracy, or geofeed authoring best practices. Helps create, refine, and improve CSV-format IP geolocation feeds with opinionated recommendations beyond RFC 8805 compliance. Do NOT use for private or internal IP address management — applies only to publicly routable IP addresses.

技能
github

Product Analytics Setup

99

How to actually instrument product analytics correctly. Event taxonomy, property design, naming conventions, schema versioning, identity stitching, funnel design, retention cohorts, North Star metric selection, dashboard hygiene, instrumentation debt, and the failure modes that produce data nobody trusts. Triggers on product analytics setup, event taxonomy, tracking plan, instrumentation, schema versioning, North Star metric, retention cohorts, funnel design, naming conventions, instrument new feature, audit existing analytics, dashboard reconciliation, instrumentation debt, Mixpanel setup, Amplitude setup, PostHog setup, warehouse-native analytics. Also triggers when the team has data but cannot trust it, or when designing instrumentation for a new feature, or when auditing an existing setup that has drifted.

技能
rampstackco

CRM Hygiene

99

Audit and improve CRM data quality by identifying missing fields, inconsistent values, duplicate records, and stale data

技能
guia-matthieu

Analytics Tracking

99

Set up, audit, and debug analytics tracking implementation — GA4, Google Tag Manager, event taxonomy, conversion tracking, and data quality. Use when building a tracking plan from scratch, auditing existing analytics for gaps or errors, debugging missing events, or setting up GTM. Trigger keywords: GA4 setup, Google Tag Manager, GTM, event tracking, analytics implementation, conversion tracking, tracking plan, event taxonomy, custom dimensions, UTM tracking, analytics audit, missing events, tracking broken. NOT for analyzing marketing campaign data — use campaign-analytics for that. NOT for BI dashboards — use product-analytics for in-product event analysis.

技能
alirezarezvani