跳转到主要内容
此内容尚未提供您的语言版本,正在以英文显示。

Implementing Llms Litgpt

技能 已验证 活跃

Implements and trains LLMs using Lightning AI's LitGPT with 20+ pretrained architectures (Llama, Gemma, Phi, Qwen, Mistral). Use when need clean model implementations, educational understanding of architectures, or production fine-tuning with LoRA/QLoRA. Single-file implementations, no abstraction layers.

目的

To enable users to easily implement, train, and fine-tune LLMs using LitGPT, offering clean code for educational understanding and production-ready workflows for advanced ML research and development.

功能

  • Implement and train LLMs with LitGPT
  • Utilize 20+ pretrained architectures (Llama, Gemma, Phi, etc.)
  • Production-ready fine-tuning with LoRA/QLoRA
  • Pretrain new models from scratch
  • Deploy models via API

使用场景

  • When needing clean, educational understanding of LLM architectures
  • For production fine-tuning with efficient methods like LoRA/QLoRA
  • When prototyping new model ideas or adapting existing architectures
  • To leverage a unified framework for various LLM training tasks

非目标

  • Providing abstraction layers beyond clean model implementations
  • Supporting every possible LLM architecture not covered by LitGPT
  • Complex, multi-agent research orchestration (handled by other skills)

安装

请先添加 Marketplace

/plugin marketplace add Orchestra-Research/AI-Research-SKILLs
/plugin install AI-Research-SKILLs@ai-research-skills

质量评分

已验证
98 /100
1 day ago 分析

信任信号

最近提交17 days ago
星标8.3k
许可证MIT
状态
查看源代码

类似扩展

Implementing Llms Litgpt

100

Implements and trains LLMs using Lightning AI's LitGPT with 20+ pretrained architectures (Llama, Gemma, Phi, Qwen, Mistral). Use when need clean model implementations, educational understanding of architectures, or production fine-tuning with LoRA/QLoRA. Single-file implementations, no abstraction layers.

技能
davila7

Peft Fine Tuning

99

Parameter-efficient fine-tuning for LLMs using LoRA, QLoRA, and 25+ methods. Use when fine-tuning large models (7B-70B) with limited GPU memory, when you need to train <1% of parameters with minimal accuracy loss, or for multi-adapter serving. HuggingFace's official library integrated with transformers ecosystem.

技能
Orchestra-Research

Unsloth

98

Expert guidance for fast fine-tuning with Unsloth - 2-5x faster training, 50-80% less memory, LoRA/QLoRA optimization

技能
Orchestra-Research

Fine Tuning Expert

98

Use when fine-tuning LLMs, training custom models, or adapting foundation models for specific tasks. Invoke for configuring LoRA/QLoRA adapters, preparing JSONL training datasets, setting hyperparameters for fine-tuning runs, adapter training, transfer learning, finetuning with Hugging Face PEFT, OpenAI fine-tuning, instruction tuning, RLHF, DPO, or quantizing and deploying fine-tuned models. Trigger terms include: LoRA, QLoRA, PEFT, finetuning, fine-tuning, adapter tuning, LLM training, model training, custom model.

技能
jeffallan

PEFT Fine Tuning

96

Parameter-efficient fine-tuning for LLMs using LoRA, QLoRA, and 25+ methods. Use when fine-tuning large models (7B-70B) with limited GPU memory, when you need to train <1% of parameters with minimal accuracy loss, or for multi-adapter serving. HuggingFace's official library integrated with transformers ecosystem.

技能
davila7

OpenVLA OFT Fine Tuning and Evaluation

98

Fine-tunes and evaluates OpenVLA-OFT and OpenVLA-OFT+ policies for robot action generation with continuous action heads, LoRA adaptation, and FiLM conditioning on LIBERO simulation and ALOHA real-world setups. Use when reproducing OpenVLA-OFT paper results, training custom VLA action heads (L1 or diffusion), deploying server-client inference for ALOHA, or debugging normalization, LoRA merge, and cross-GPU issues.

技能
Orchestra-Research