跳转到主要内容
此内容尚未提供您的语言版本,正在以英文显示。

Weights And Biases

技能 活跃

Track ML experiments with automatic logging, visualize training in real-time, optimize hyperparameters with sweeps, and manage model registry with W&B - collaborative MLOps platform

目的

To streamline and automate the process of tracking, visualizing, and optimizing machine learning experiments using the Weights & Biases platform.

功能

  • Automatic ML experiment tracking
  • Real-time training visualization
  • Hyperparameter optimization with sweeps
  • Model registry management
  • Data and model versioning with artifacts

使用场景

  • Use when you need to log metrics and track configurations for ML experiments.
  • Use to visualize model training progress and compare different runs.
  • Use to automate hyperparameter tuning to find optimal model configurations.
  • Use to manage and version ML models and datasets throughout the MLOps lifecycle.

非目标

  • This skill does not replace the core functionality of the Weights & Biases platform itself.
  • It does not provide generic ML model training or data preprocessing capabilities outside of W&B integration.

Trust

  • warning:Issues AttentionIn the last 90 days, 17 issues were opened and 4 were closed, indicating a low closure rate and potential maintainer responsiveness issues.

Execution

  • info:Pinned dependenciesThe skill relies on the 'wandb' package. While the SKILL.md specifies it as a dependency, a lockfile or specific version pinning is not explicitly shown for the 'wandb' package itself within the skill's bundle.

安装

npx skills add davila7/claude-code-templates

通过 npx 运行 Vercel skills CLI(skills.sh)— 需要本地安装 Node.js,以及至少一个兼容 skills 的智能体(Claude Code、Cursor、Codex 等)。前提是仓库遵循 agentskills.io 格式。

质量评分

95 /100
1 day ago 分析

信任信号

最近提交1 day ago
星标27.2k
许可证MIT
状态
查看源代码

类似扩展

Weights And Biases

96

Track ML experiments with automatic logging, visualize training in real-time, optimize hyperparameters with sweeps, and manage model registry with W&B - collaborative MLOps platform

技能
Orchestra-Research

Mlflow

98

Track ML experiments, manage model registry with versioning, deploy models to production, and reproduce experiments with MLflow - framework-agnostic ML lifecycle platform

技能
davila7

MLflow

96

Track ML experiments, manage model registry with versioning, deploy models to production, and reproduce experiments with MLflow - framework-agnostic ML lifecycle platform

技能
Orchestra-Research

Hf Cli

100

Hugging Face Hub CLI (`hf`) for downloading, uploading, and managing models, datasets, spaces, buckets, repos, papers, jobs, and more on the Hugging Face Hub. Use when: handling authentication; managing local cache; managing Hugging Face Buckets; running or scheduling jobs on Hugging Face infrastructure; managing Hugging Face repos; discussions and pull requests; browsing models, datasets and spaces; reading, searching, or browsing academic papers; managing collections; querying datasets; configuring spaces; setting up webhooks; or deploying and managing HF Inference Endpoints. Make sure to use this skill whenever the user mentions 'hf', 'huggingface', 'Hugging Face', 'huggingface-cli', or 'hugging face cli', or wants to do anything related to the Hugging Face ecosystem and to AI and ML in general. Also use for cloud storage needs like training checkpoints, data pipelines, or agent traces. Use even if the user doesn't explicitly ask for a CLI command. Replaces the deprecated `huggingface-cli`.

技能
huggingface

Arize Experiment

100

Creates, runs, and analyzes Arize experiments for evaluating and comparing model performance. Covers experiment CRUD, exporting runs, comparing results, and evaluation workflows using the ax CLI. Use when the user mentions create experiment, run experiment, compare models, model performance, evaluate AI, experiment results, benchmark, A/B test models, or measure accuracy.

技能
github

Arize Evaluator

100

Handles LLM-as-judge evaluation workflows on Arize including creating/updating evaluators, running evaluations on spans or experiments, managing tasks, trigger-run operations, column mapping, and continuous monitoring. Use when the user mentions create evaluator, LLM judge, hallucination, faithfulness, correctness, relevance, run eval, score spans, score experiment, trigger-run, column mapping, continuous monitoring, or improve evaluator prompt.

技能
github