Axolotl
技能 已验证 活跃Expert guidance for fine-tuning LLMs with Axolotl - YAML configs, 100+ models, LoRA/QLoRA, DPO/KTO/ORPO/GRPO, multimodal support
To serve as an expert resource for users fine-tuning LLMs with Axolotl, offering detailed configuration guidance, best practices, and troubleshooting information.
功能
- Expert guidance for Axolotl fine-tuning
- Detailed YAML configuration explanations
- Support for 100+ LLM models
- Coverage of LoRA, QLoRA, DPO, KTO, ORPO, GRPO tuning methods
- Information on multimodal support
使用场景
- Fine-tuning LLMs with Axolotl
- Understanding Axolotl features and APIs
- Implementing Axolotl solutions
- Debugging Axolotl code and configurations
非目标
- Providing generic LLM fine-tuning advice outside of Axolotl
- Acting as a direct interface to run Axolotl commands (it provides guidance only)
安装
请先添加 Marketplace
/plugin marketplace add Orchestra-Research/AI-Research-SKILLs/plugin install AI-Research-SKILLs@ai-research-skills质量评分
已验证类似扩展
Axolotl Fine Tuning Skill
98Expert guidance for fine-tuning LLMs with Axolotl - YAML configs, 100+ models, LoRA/QLoRA, DPO/KTO/ORPO/GRPO, multimodal support
Implementing Llms Litgpt
100Implements and trains LLMs using Lightning AI's LitGPT with 20+ pretrained architectures (Llama, Gemma, Phi, Qwen, Mistral). Use when need clean model implementations, educational understanding of architectures, or production fine-tuning with LoRA/QLoRA. Single-file implementations, no abstraction layers.
Unsloth
100Expert guidance for fast fine-tuning with Unsloth - 2-5x faster training, 50-80% less memory, LoRA/QLoRA optimization
Train Sentence Transformers
98Train or fine-tune sentence-transformers models across `SentenceTransformer` (bi-encoder; dense or static embedding model; for retrieval, similarity, clustering, classification, paraphrase mining, dedup, multimodal), `CrossEncoder` (reranker; pair scoring for two-stage retrieval / pair classification), and `SparseEncoder` (SPLADE, sparse embedding model; for learned-sparse retrieval). Covers loss selection, hard-negative mining, evaluators, distillation, LoRA, Matryoshka, and Hugging Face Hub publishing. Use for any sentence-transformers training task.
PyTorch Lightning
100Deep learning framework (PyTorch Lightning). Organize PyTorch code into LightningModules, configure Trainers for multi-GPU/TPU, implement data pipelines, callbacks, logging (W&B, TensorBoard), distributed training (DDP, FSDP, DeepSpeed), for scalable neural network training.
Chat Format
100Format prompts for different LLM providers with chat templates and HNSW-powered context retrieval