Instructor
技能 活跃Extract structured data from LLM responses with Pydantic validation, retry failed extractions automatically, parse complex JSON with type safety, and stream partial results with Instructor - battle-tested structured output library
To reliably extract and validate structured data from LLM responses, enabling safer and more predictable integration of LLM outputs into applications.
功能
- Extract structured data with Pydantic models
- Automatic validation of LLM outputs
- Retry failed extractions with error feedback
- Parse complex JSON with type safety
- Stream partial results for real-time processing
使用场景
- Extracting user profiles, product details, or financial data from text
- Classifying text into predefined categories with confidence scores
- Parsing complex nested JSON outputs from LLMs
- Building real-time applications that consume LLM-generated structured data
非目标
- Replacing the core LLM provider itself
- Performing complex data transformations beyond validation
- Handling LLM API calls directly without Instructor's structured output features
Trust
- warning:Issues AttentionThere are 17 open issues and 4 closed issues in the last 90 days, indicating a low closure rate and potentially slow maintainer response.
安装
npx skills add davila7/claude-code-templates通过 npx 运行 Vercel skills CLI(skills.sh)— 需要本地安装 Node.js,以及至少一个兼容 skills 的智能体(Claude Code、Cursor、Codex 等)。前提是仓库遵循 agentskills.io 格式。
质量评分
类似扩展
Instructor
98Extract structured data from LLM responses with Pydantic validation, retry failed extractions automatically, parse complex JSON with type safety, and stream partial results with Instructor - battle-tested structured output library
Arize Prompt Optimization
100Optimizes, improves, and debugs LLM prompts using production trace data, evaluations, and annotations. Extracts prompts from spans, gathers performance signal, and runs a data-driven optimization loop using the ax CLI. Use when the user mentions optimize prompt, improve prompt, make AI respond better, improve output quality, prompt engineering, prompt tuning, or system prompt improvement.
Prompt Optimization
100应用提示重复以提高非推理 LLM 的准确性
创建原子工具
99构建一个 `BaseTool[InSchema, OutSchema]` 子类 — 输入/输出模式,`BaseToolConfig`,`run()`(和可选的 `run_async()`),环境变量驱动的 secret,类型化的失败输出。当用户要求“添加工具”、“创建工具”、“将 API 包装成工具”、“构建 `BaseTool`”、“制作计算器/搜索/天气工具”或运行 `/atomic-agents:create-atomic-tool` 时使用。
Guidance
99Control LLM output with regex and grammars, guarantee valid JSON/XML/code generation, enforce structured formats, and build multi-step workflows with Guidance - Microsoft Research's constrained generation framework
Create Atomic Schema
98设计和编写 Atomic Agents 代理或工具的 `BaseIOSchema` 输入/输出对 — 文档字符串、字段描述、验证器、错误变体。当用户要求“创建 schema”、“设计输入/输出 schema”、“定义 `IOSchema`”、“编写 `BaseIOSchema`”、“建模代理的输出”或运行 `/atomic-agents:create-atomic-schema` 时使用。