跳转到主要内容
此内容尚未提供您的语言版本,正在以英文显示。

Arize Prompt Optimization

技能 已验证 活跃

Optimizes, improves, and debugs LLM prompts using production trace data, evaluations, and annotations. Extracts prompts from spans, gathers performance signal, and runs a data-driven optimization loop using the ax CLI. Use when the user mentions optimize prompt, improve prompt, make AI respond better, improve output quality, prompt engineering, prompt tuning, or system prompt improvement.

目的

To enable users to systematically improve LLM prompt performance by analyzing production trace data and applying a data-driven optimization process.

功能

  • Optimize LLM prompts using production trace data
  • Extract prompts from spans
  • Gather LLM performance signals
  • Run data-driven optimization loops with `ax` CLI
  • Debug and improve LLM output quality

使用场景

  • When needing to improve AI response quality
  • For prompt engineering and tuning
  • When improving system prompts based on performance metrics
  • When analyzing LLM output for correctness and faithfulness

非目标

  • Directly modifying LLM models
  • Collecting trace data (relies on external instrumentation)
  • Managing Arize platform infrastructure

工作流

  1. Extract the current prompt from production trace data
  2. Gather performance data from traces, datasets, and experiments
  3. Analyze failures and identify patterns for optimization
  4. Generate a revised prompt using a meta-prompt template
  5. Apply the revised prompt and iterate on the optimization loop

先决条件

  • Requires the `ax` CLI
  • Requires a configured Arize profile

安装

npx skills add github/awesome-copilot

通过 npx 运行 Vercel skills CLI(skills.sh)— 需要本地安装 Node.js,以及至少一个兼容 skills 的智能体(Claude Code、Cursor、Codex 等)。前提是仓库遵循 agentskills.io 格式。

质量评分

已验证
100 /100
1 day ago 分析

信任信号

最近提交1 day ago
星标32.9k
许可证MIT
状态
查看源代码

类似扩展

Prompt Optimization

100

应用提示重复以提高非推理 LLM 的准确性

技能
asklokesh

Arize Ai Provider Integration

100

Creates, reads, updates, and deletes Arize AI integrations that store LLM provider credentials used by evaluators and other Arize features. Supports any LLM provider (e.g. OpenAI, Anthropic, Azure OpenAI, AWS Bedrock, Vertex AI, Gemini, NVIDIA NIM). Use when the user mentions AI integration, LLM provider credentials, create integration, list integrations, update credentials, delete integration, or connecting an LLM provider to Arize.

技能
github

Oh My Claudecode

100

Process-first advisor routing for Claude, Codex, or Gemini via `omc ask`, with artifact capture and no raw CLI assembly

技能
Yeachan-Heo

Unsloth

100

Expert guidance for fast fine-tuning with Unsloth - 2-5x faster training, 50-80% less memory, LoRA/QLoRA optimization

技能
davila7

CE Optimize

100

Run metric-driven iterative optimization loops -- define a measurable goal, run parallel experiments, measure each against hard gates or LLM-as-judge scores, keep improvements, and converge on the best solution. Use when optimizing clustering quality, search relevance, build performance, prompt quality, or any measurable outcome that benefits from systematic experimentation.

技能
EveryInc

Arize Experiment

100

Creates, runs, and analyzes Arize experiments for evaluating and comparing model performance. Covers experiment CRUD, exporting runs, comparing results, and evaluation workflows using the ax CLI. Use when the user mentions create experiment, run experiment, compare models, model performance, evaluate AI, experiment results, benchmark, A/B test models, or measure accuracy.

技能
github