Zum Hauptinhalt springen
Dieser Inhalt ist noch nicht in Ihrer Sprache verfügbar und wird auf Englisch angezeigt.

Fine Tuning Expert

Skill Verifiziert Aktiv

Use when fine-tuning LLMs, training custom models, or adapting foundation models for specific tasks. Invoke for configuring LoRA/QLoRA adapters, preparing JSONL training datasets, setting hyperparameters for fine-tuning runs, adapter training, transfer learning, finetuning with Hugging Face PEFT, OpenAI fine-tuning, instruction tuning, RLHF, DPO, or quantizing and deploying fine-tuned models. Trigger terms include: LoRA, QLoRA, PEFT, finetuning, fine-tuning, adapter tuning, LLM training, model training, custom model.

Zweck

To serve as an expert resource for anyone fine-tuning LLMs, offering best practices, code examples, and configuration guidance for various stages of the process.

Funktionen

  • Detailed LoRA/QLoRA configuration guidance
  • Dataset preparation and validation utilities
  • Hyperparameter tuning strategies and examples
  • Code examples for training and deployment
  • Best practices for PEFT and model optimization

Anwendungsfälle

  • Configuring LoRA/QLoRA adapters for custom LLM tasks
  • Preparing and validating JSONL training datasets
  • Setting hyperparameters for fine-tuning runs
  • Optimizing fine-tuned models for deployment
  • Adapting foundation models with PEFT methods

Nicht-Ziele

  • Performing the fine-tuning process itself (provides guidance, not execution)
  • Training foundation models from scratch
  • Managing cloud infrastructure for training
  • Handling non-ML related tasks

Installation

Zuerst Marketplace hinzufügen

/plugin marketplace add jeffallan/claude-skills
/plugin install claude-skills@fullstack-dev-skills

Qualitätspunktzahl

Verifiziert
98 /100
Analysiert about 22 hours ago

Vertrauenssignale

Letzter Commit13 days ago
Sterne9k
LizenzMIT
Status
Quellcode ansehen

Ähnliche Erweiterungen

Peft Fine Tuning

99

Parameter-efficient fine-tuning for LLMs using LoRA, QLoRA, and 25+ methods. Use when fine-tuning large models (7B-70B) with limited GPU memory, when you need to train <1% of parameters with minimal accuracy loss, or for multi-adapter serving. HuggingFace's official library integrated with transformers ecosystem.

Skill
Orchestra-Research

PEFT Fine Tuning

96

Parameter-efficient fine-tuning for LLMs using LoRA, QLoRA, and 25+ methods. Use when fine-tuning large models (7B-70B) with limited GPU memory, when you need to train <1% of parameters with minimal accuracy loss, or for multi-adapter serving. HuggingFace's official library integrated with transformers ecosystem.

Skill
davila7

Unsloth

98

Expert guidance for fast fine-tuning with Unsloth - 2-5x faster training, 50-80% less memory, LoRA/QLoRA optimization

Skill
Orchestra-Research

Implementing Llms Litgpt

98

Implements and trains LLMs using Lightning AI's LitGPT with 20+ pretrained architectures (Llama, Gemma, Phi, Qwen, Mistral). Use when need clean model implementations, educational understanding of architectures, or production fine-tuning with LoRA/QLoRA. Single-file implementations, no abstraction layers.

Skill
Orchestra-Research

Implementing Llms Litgpt

100

Implements and trains LLMs using Lightning AI's LitGPT with 20+ pretrained architectures (Llama, Gemma, Phi, Qwen, Mistral). Use when need clean model implementations, educational understanding of architectures, or production fine-tuning with LoRA/QLoRA. Single-file implementations, no abstraction layers.

Skill
davila7

Unsloth

100

Expert guidance for fast fine-tuning with Unsloth - 2-5x faster training, 50-80% less memory, LoRA/QLoRA optimization

Skill
davila7