跳转到主要内容
此内容尚未提供您的语言版本,正在以英文显示。

Knowledge Distillation

技能 已验证 活跃

Compress large language models using knowledge distillation from teacher to student models. Use when deploying smaller models with retained performance, transferring GPT-4 capabilities to open-source models, or reducing inference costs. Covers temperature scaling, soft targets, reverse KLD, logit distillation, and MiniLLM training strategies.

目的

Compress large language models using knowledge distillation from teacher to student models, enabling the deployment of smaller, high-performing models and reducing inference costs.

功能

  • Compress LLMs using knowledge distillation
  • Transfer capabilities from large to open-source models
  • Implement temperature scaling and soft targets
  • Utilize MiniLLM (Reverse KLD) for generative models
  • Perform response distillation via synthetic data

使用场景

  • Compressing models from 70B to 7B while retaining performance
  • Transferring capabilities from proprietary models like GPT-4 to open-source models
  • Reducing inference costs by deploying smaller student models
  • Creating specialized models by distilling domain-specific knowledge

非目标

  • Training LLMs from scratch
  • Developing new model architectures
  • Evaluating LLM performance on tasks unrelated to distillation

Code Execution

  • info:LoggingThe `transformers` library and standard Python logging are used, but a dedicated audit log file for destructive actions is not explicitly mentioned or implemented within the skill's scope.

Execution

  • info:Pinned dependenciesDependencies are listed, but lockfiles are not explicitly mentioned in the documentation, and scripts lack detailed shebangs/headers.

安装

请先添加 Marketplace

/plugin marketplace add Orchestra-Research/AI-Research-SKILLs
/plugin install AI-Research-SKILLs@ai-research-skills

质量评分

已验证
98 /100
1 day ago 分析

信任信号

最近提交17 days ago
星标8.3k
许可证MIT
状态
查看源代码

类似扩展

Knowledge Distillation

96

Compress large language models using knowledge distillation from teacher to student models. Use when deploying smaller models with retained performance, transferring GPT-4 capabilities to open-source models, or reducing inference costs. Covers temperature scaling, soft targets, reverse KLD, logit distillation, and MiniLLM training strategies.

技能
davila7

Chat Format

100

Format prompts for different LLM providers with chat templates and HNSW-powered context retrieval

技能
ruvnet

Oh My Claudecode

100

Process-first advisor routing for Claude, Codex, or Gemini via `omc ask`, with artifact capture and no raw CLI assembly

技能
Yeachan-Heo

Wrap Up Ritual

100

End-of-session ritual that audits changes, runs quality checks, captures learnings, and produces a session summary. Use when saying "wrap up", "done for the day", "finish coding", or ending a coding session.

技能
rohitg00

Project Development

100

This skill should be used when the user asks to "start an LLM project", "design batch pipeline", "evaluate task-model fit", "structure agent project", or mentions pipeline architecture, agent-assisted development, cost estimation, or choosing between LLM and traditional approaches.

技能
muratcankoylan

Context Compression

100

This skill should be used when the user asks to "compress context", "summarize conversation history", "implement compaction", "reduce token usage", or mentions context compression, structured summarization, tokens-per-task optimization, or long-running agent sessions exceeding context limits.

技能
muratcankoylan