Zum Hauptinhalt springen
Dieser Inhalt ist noch nicht in Ihrer Sprache verfügbar und wird auf Englisch angezeigt.

Data Quality Frameworks

Skill Verifiziert Aktiv

Implement data quality validation with Great Expectations, dbt tests, and data contracts. Use when building data quality pipelines, implementing validation rules, or establishing data contracts.

Zweck

Implement comprehensive data quality validation and testing within data pipelines using industry-standard tools like Great Expectations, dbt, and data contracts.

Funktionen

  • Implement data quality validation with Great Expectations
  • Build dbt test suites
  • Establish data contracts
  • Automate data validation pipelines
  • Monitor data quality metrics

Anwendungsfälle

  • Building data quality pipelines
  • Implementing validation rules
  • Establishing data contracts between teams
  • Monitoring data quality metrics

Nicht-Ziele

  • Full data pipeline orchestration beyond quality checks
  • Development of the data quality tools themselves
  • Data transformation or ETL logic

Workflow

  1. Understand data quality dimensions and testing pyramid.
  2. Set up Great Expectations (installation, datasource, expectations).
  3. Configure Great Expectations checkpoints for validation.
  4. Implement dbt tests for data models.
  5. Define and implement data contracts.
  6. Automate quality checks within a pipeline.

Praktiken

  • Data Quality
  • Testing
  • Pipeline Development

Voraussetzungen

  • Python environment
  • Great Expectations installed
  • dbt installed

Installation

Zuerst Marketplace hinzufügen

/plugin marketplace add wshobson/agents
/plugin install data-engineering@claude-code-workflows

Qualitätspunktzahl

Verifiziert
97 /100
Analysiert about 17 hours ago

Vertrauenssignale

Letzter Commit3 days ago
Sterne35.3k
LizenzMIT
Status
Quellcode ansehen

Ähnliche Erweiterungen

Data Quality Auditor

97

Audit datasets for completeness, consistency, accuracy, and validity. Profile data distributions, detect anomalies and outliers, surface structural issues, and produce an actionable remediation plan.

Skill
alirezarezvani

Senior Data Engineer

95

Data engineering skill for building scalable data pipelines, ETL/ELT systems, and data infrastructure. Expertise in Python, SQL, Spark, Airflow, dbt, Kafka, and modern data stack. Includes data modeling, pipeline orchestration, data quality, and DataOps. Use when designing data architectures, building data pipelines, optimizing data workflows, implementing data governance, or troubleshooting data issues.

Skill
alirezarezvani

Geofeed Tuner

100

Use this skill whenever the user mentions IP geolocation feeds, RFC 8805, geofeeds, or wants help creating, tuning, validating, or publishing a self-published IP geolocation feed in CSV format. Intended user audience is a network operator, ISP, mobile carrier, cloud provider, hosting company, IXP, or satellite provider asking about IP geolocation accuracy, or geofeed authoring best practices. Helps create, refine, and improve CSV-format IP geolocation feeds with opinionated recommendations beyond RFC 8805 compliance. Do NOT use for private or internal IP address management — applies only to publicly routable IP addresses.

Skill
github

Product Analytics Setup

99

How to actually instrument product analytics correctly. Event taxonomy, property design, naming conventions, schema versioning, identity stitching, funnel design, retention cohorts, North Star metric selection, dashboard hygiene, instrumentation debt, and the failure modes that produce data nobody trusts. Triggers on product analytics setup, event taxonomy, tracking plan, instrumentation, schema versioning, North Star metric, retention cohorts, funnel design, naming conventions, instrument new feature, audit existing analytics, dashboard reconciliation, instrumentation debt, Mixpanel setup, Amplitude setup, PostHog setup, warehouse-native analytics. Also triggers when the team has data but cannot trust it, or when designing instrumentation for a new feature, or when auditing an existing setup that has drifted.

Skill
rampstackco

CRM Hygiene

99

Audit and improve CRM data quality by identifying missing fields, inconsistent values, duplicate records, and stale data

Skill
guia-matthieu

Analytics Tracking

99

Set up, audit, and debug analytics tracking implementation — GA4, Google Tag Manager, event taxonomy, conversion tracking, and data quality. Use when building a tracking plan from scratch, auditing existing analytics for gaps or errors, debugging missing events, or setting up GTM. Trigger keywords: GA4 setup, Google Tag Manager, GTM, event tracking, analytics implementation, conversion tracking, tracking plan, event taxonomy, custom dimensions, UTM tracking, analytics audit, missing events, tracking broken. NOT for analyzing marketing campaign data — use campaign-analytics for that. NOT for BI dashboards — use product-analytics for in-product event analysis.

Skill
alirezarezvani