Chdb Sql
Skill Verified ActiveIn-process ClickHouse SQL engine for Python — run ClickHouse SQL queries directly on local files, remote databases, and cloud storage without a server. Use when the user wants to write SQL queries against Parquet/CSV/ JSON files, use ClickHouse table functions (mysql(), s3(), postgresql(), iceberg(), deltaLake() etc.), build stateful analytical pipelines with Session, use parametrized queries, window functions, or other advanced ClickHouse SQL features. Also use when the user explicitly mentions chdb.query(), ClickHouse SQL syntax, or wants cross-source SQL joins. Do NOT use for pandas-style DataFrame operations — use chdb-datastore instead.
To enable users to leverage the power of ClickHouse SQL directly within their Python environment for efficient data querying and analysis across various data sources.
Features
- Run ClickHouse SQL queries in-process
- Query local files (Parquet, CSV, JSON)
- Connect to remote databases (MySQL, PostgreSQL)
- Access cloud storage (S3, GCS, Azure Blob)
- Use advanced ClickHouse SQL features (table functions, window functions, parametrized queries)
- Integrate Python data structures (dicts, DataFrames) as SQL tables
- Build stateful analytical pipelines with Session API
Use Cases
- Querying data from Parquet, CSV, or JSON files using SQL.
- Joining data from different sources (e.g., MySQL and S3) with SQL.
- Building multi-step analytical pipelines with persistent sessions.
- Leveraging ClickHouse's rich SQL dialect within a Python script.
Non-Goals
- Performing pandas-style DataFrame operations (use `chdb-datastore` instead).
- Running a standalone ClickHouse server.
- Acting as a general-purpose database client for all SQL dialects.
Trust
- info:Issues AttentionOpen issues: 2, Closed issues: 0. The maintainers have not closed any issues in the last 90 days, and there are currently 2 open issues.
Installation
First, add the marketplace
/plugin marketplace add clickhouse/agent-skills/plugin install agent-skills@clickhouse-agent-skillsQuality Score
VerifiedTrust Signals
Similar Extensions
Fit Drift Diffusion Model
100Fit cognitive drift-diffusion models (Ratcliff DDM) to reaction time and accuracy data with parameter estimation (drift rate, boundary separation, non-decision time), model comparison, and parameter recovery validation. Use when modeling binary decision-making with reaction time data, estimating cognitive parameters from experimental data, comparing sequential sampling model variants, or decomposing speed-accuracy tradeoff effects into latent cognitive components.
DBHub Database Query Guide
100Guide for querying databases through DBHub MCP server. Use this skill whenever you need to explore database schemas, inspect tables, or run SQL queries via DBHub's MCP tools (search_objects, execute_sql). Activates on any database query task, schema exploration, data retrieval, or SQL execution through MCP — even if the user just says "check the database" or "find me some data." This skill ensures you follow the correct explore-first workflow instead of guessing table structures.
Dsql
100Build with Aurora DSQL — manage schemas, execute queries, handle migrations, diagnose query plans, and develop applications with a serverless, distributed SQL database. Covers IAM auth, multi-tenant patterns, MySQL-to-DSQL migration, DDL operations, query plan explainability, and SQL compatibility validation. Triggers on phrases like: DSQL, Aurora DSQL, create DSQL table, DSQL schema, migrate to DSQL, distributed SQL database, serverless PostgreSQL-compatible database, DSQL query plan, DSQL EXPLAIN ANALYZE, why is my DSQL query slow.
MySQL Read Only Query Skill
99Execute read-only SQL queries against multiple MySQL databases. Use when: (1) querying MySQL databases, (2) exploring database schemas/tables, (3) running SELECT queries for data analysis, (4) checking database contents. Supports multiple database connections with descriptions for intelligent auto-selection. Blocks all write operations (INSERT, UPDATE, DELETE, DROP, etc.) for safety.
Serialize Data Formats
97Serialize and deserialize data across common formats including JSON, XML, YAML, Protocol Buffers, MessagePack, and Apache Arrow/Parquet. Covers format selection criteria, encoding/decoding patterns, performance trade-offs, and interoperability considerations. Use when choosing a wire format for API communication, persisting structured data to disk, exchanging data between systems written in different languages, optimizing transfer size or parsing speed, or migrating from one serialization format to another.
ClickHouse Best Practices
88MUST USE when reviewing ClickHouse schemas, queries, or configurations. Contains 31 rules that MUST be checked before providing recommendations. Always read relevant rule files and cite specific rules in responses.