跳转到主要内容
此内容尚未提供您的语言版本,正在以英文显示。

Huggingface Llm Trainer

技能 已验证 活跃

Train or fine-tune language and vision models using TRL (Transformer Reinforcement Learning) or Unsloth with Hugging Face Jobs infrastructure. Covers SFT, DPO, GRPO and reward modeling training methods, plus GGUF conversion for local deployment. Includes guidance on the TRL Jobs package, UV scripts with PEP 723 format, dataset preparation and validation, hardware selection, cost estimation, Trackio monitoring, Hub authentication, model selection/leaderboards and model persistence. Use for tasks involving cloud GPU training, GGUF conversion, or when users mention training on Hugging Face Jobs without local GPU setup.

目的

Streamline and simplify the process of training and converting LLMs on cloud infrastructure, making advanced ML workflows accessible.

功能

  • Fine-tune LLMs using TRL or Unsloth
  • Leverage Hugging Face Jobs infrastructure
  • Supports SFT, DPO, GRPO, and Reward Modeling
  • Convert models to GGUF format for local deployment
  • Includes cost estimation and Trackio monitoring

使用场景

  • Fine-tune language models on cloud GPUs without local setup
  • Align models with human preferences using DPO
  • Convert trained models to GGUF for Ollama or LM Studio
  • Optimize training for limited GPU memory with Unsloth

非目标

  • Directly managing Hugging Face infrastructure (handled by `hf-cli`)
  • Advanced distributed training setup beyond TRL's automatic handling
  • Modifying the core TRL or Unsloth libraries

安装

/plugin install skills@huggingface-skills

质量评分

已验证
99 /100
1 day ago 分析

信任信号

最近提交2 days ago
星标10.5k
许可证Apache-2.0
状态
查看源代码

类似扩展