Dask Data Science
Skill Verifiziert AktivPart of the AlterLab Academic Skills suite. Distributed computing for larger-than-RAM pandas/NumPy workflows. Use when you need to scale existing pandas/NumPy code beyond memory or across clusters. Best for parallel file processing, distributed ML, integration with existing pandas code. For out-of-core analytics on single machine use vaex; for in-memory speed use polars.
To provide an expert assistant for scaling data science workflows using Dask, enabling users to process datasets that exceed single-machine memory or require parallel computation.
Funktionen
- Distributed computing for pandas/NumPy
- Larger-than-memory data processing
- Parallel file processing
- Integration with existing pandas/NumPy code
- Scales from laptops to clusters
Anwendungsfälle
- Scaling pandas operations to larger datasets
- Parallelizing computations for performance
- Processing multiple files efficiently (CSVs, Parquet, JSON)
- Distributing workloads across multiple cores or machines
Nicht-Ziele
- Out-of-core analytics on a single machine (use vaex)
- In-memory speed optimization (use polars)
- Replacing core pandas/NumPy functionality for in-memory data
Workflow
- Load data using Dask's parallel readers (read_csv, read_parquet)
- Perform operations (filtering, transformations, aggregations) on Dask DataFrames, Arrays, or Bags
- Leverage Dask's lazy evaluation and task graph construction
- Trigger computation with .compute() or dask.compute()
- Optimize performance through chunking, persist, and scheduler selection
- Save results or convert to pandas for final analysis
Installation
npx skills add AlterLab-IEU/AlterLab-Academic-SkillsFührt das Vercel skills CLI (skills.sh) via npx aus — benötigt Node.js lokal und mindestens einen installierten skills-kompatiblen Agent (Claude Code, Cursor, Codex, …). Setzt voraus, dass das Repo dem agentskills.io-Format folgt.
Qualitätspunktzahl
VerifiziertVertrauenssignale
Ähnliche Erweiterungen
AlterLab Zarr
99Part of the AlterLab Academic Skills suite. Chunked N-D arrays for cloud storage. Compressed arrays, parallel I/O, S3/GCS integration, NumPy/Dask/Xarray compatible, for large-scale scientific computing pipelines.
Dask
98Distributed computing for larger-than-RAM pandas/NumPy workflows. Use when you need to scale existing pandas/NumPy code beyond memory or across clusters. Best for parallel file processing, distributed ML, integration with existing pandas code. For out-of-core analytics on single machine use vaex; for in-memory speed use polars.
Spark Engineer
99Use when writing Spark jobs, debugging performance issues, or configuring cluster settings for Apache Spark applications, distributed data processing pipelines, or big data workloads. Invoke to write DataFrame transformations, optimize Spark SQL queries, implement RDD pipelines, tune shuffle operations, configure executor memory, process .parquet files, handle data partitioning, or build structured streaming analytics.
Zarr Python
97Chunked N-D arrays for cloud storage. Compressed arrays, parallel I/O, S3/GCS integration, NumPy/Dask/Xarray compatible, for large-scale scientific computing pipelines.
OraClaw Forecast
100Zeitreihenprognose für KI-Agenten. ARIMA- und Holt-Winters-Vorhersagen mit Konfidenzintervallen. Prognostizieren Sie Umsatz, Traffic, Preise oder beliebige sequentielle Daten. Inferenz unter 5 ms.
SHAP Model Interpretability
100Model interpretability and explainability using SHAP (SHapley Additive exPlanations). Use this skill when explaining machine learning model predictions, computing feature importance, generating SHAP plots (waterfall, beeswarm, bar, scatter, force, heatmap), debugging models, analyzing model bias or fairness, comparing models, or implementing explainable AI. Works with tree-based models (XGBoost, LightGBM, Random Forest), deep learning (TensorFlow, PyTorch), linear models, and any black-box model.