Skip to main content

Unsloth

Skill Verified Active

Expert guidance for fast fine-tuning with Unsloth - 2-5x faster training, 50-80% less memory, LoRA/QLoRA optimization

Purpose

To offer expert guidance and readily accessible documentation for fast and memory-efficient fine-tuning of LLMs using the Unsloth library.

Features

  • Fast fine-tuning guidance
  • Memory optimization advice
  • LoRA/QLoRA implementation details
  • Comprehensive Unsloth documentation
  • Best practices for LLM fine-tuning

Use Cases

  • When working with the Unsloth library for LLM fine-tuning.
  • Seeking information on Unsloth features, APIs, or installation.
  • Debugging Unsloth-related code or implementation challenges.
  • Learning best practices for efficient LLM fine-tuning.

Non-Goals

  • Providing direct code execution capabilities.
  • Acting as a generic LLM fine-tuning tool outside of Unsloth.
  • Offering support for libraries other than Unsloth.

Installation

npx skills add davila7/claude-code-templates

Runs the Vercel skills CLI (skills.sh) via npx — needs Node.js locally and at least one installed skills-compatible agent (Claude Code, Cursor, Codex, …). Assumes the repo follows the agentskills.io format.

Quality Score

Verified
100 /100
Analyzed about 22 hours ago

Trust Signals

Last commitabout 24 hours ago
Stars27.2k
LicenseMIT
Status
View Source

Similar Extensions

Implementing Llms Litgpt

100

Implements and trains LLMs using Lightning AI's LitGPT with 20+ pretrained architectures (Llama, Gemma, Phi, Qwen, Mistral). Use when need clean model implementations, educational understanding of architectures, or production fine-tuning with LoRA/QLoRA. Single-file implementations, no abstraction layers.

Skill
davila7

Arize Prompt Optimization

100

Optimizes, improves, and debugs LLM prompts using production trace data, evaluations, and annotations. Extracts prompts from spans, gathers performance signal, and runs a data-driven optimization loop using the ax CLI. Use when the user mentions optimize prompt, improve prompt, make AI respond better, improve output quality, prompt engineering, prompt tuning, or system prompt improvement.

Skill
github

Prompt Optimization

100

Applies prompt repetition to improve accuracy for non-reasoning LLMs

Skill
asklokesh

Huggingface Llm Trainer

99

Train or fine-tune language and vision models using TRL (Transformer Reinforcement Learning) or Unsloth with Hugging Face Jobs infrastructure. Covers SFT, DPO, GRPO and reward modeling training methods, plus GGUF conversion for local deployment. Includes guidance on the TRL Jobs package, UV scripts with PEP 723 format, dataset preparation and validation, hardware selection, cost estimation, Trackio monitoring, Hub authentication, model selection/leaderboards and model persistence. Use for tasks involving cloud GPU training, GGUF conversion, or when users mention training on Hugging Face Jobs without local GPU setup.

Skill
huggingface

Unsloth

98

Expert guidance for fast fine-tuning with Unsloth - 2-5x faster training, 50-80% less memory, LoRA/QLoRA optimization

Skill
Orchestra-Research

Teach Guidance

100

Guide a person in becoming a better teacher and explainer. AI coaches content structuring, audience calibration, explanation clarity, Socratic questioning technique, feedback interpretation, and reflective practice for technical presentations, documentation, and mentoring. Use when a person needs to present technical content and wants preparation coaching, wants to write better documentation or tutorials, struggles to explain concepts across expertise levels, is mentoring a colleague, or is preparing for a talk or knowledge-sharing session.

Skill
pjt222

© 2025 SkillRepo · Find the right skill, skip the noise.