跳转到主要内容
此内容尚未提供您的语言版本,正在以英文显示。

Huggingface Local Models

插件 已验证 活跃

Use to select models to run locally with llama.cpp and GGUF on CPU, Mac Metal, CUDA, or ROCm. Covers finding GGUFs, quant selection, running servers, exact GGUF file lookup, conversion, and OpenAI-compatible local serving.

目的

To enable users to easily run large language models locally on their own hardware, leveraging optimized tools like llama.cpp and Hugging Face's model repository.

功能

  • Select local LLMs with llama.cpp and GGUF
  • Support for CPU, Mac Metal, CUDA, and ROCm
  • Find and select appropriate GGUF models and quantizations
  • Run local LLM servers and CLI interfaces
  • Convert models when GGUF is not directly available

使用场景

  • Running LLMs locally for privacy or cost savings.
  • Experimenting with different local LLM configurations and hardware.
  • Developing applications that require a local inference backend.

非目标

  • Providing a managed cloud LLM service.
  • Acting as a general-purpose LLM API wrapper.
  • Abstracting away the underlying llama.cpp or Hugging Face Hub tooling entirely.

安装

请先添加 Marketplace

/plugin marketplace add huggingface/skills
/plugin install huggingface-local-models@huggingface-skills

质量评分

已验证
99 /100
1 day ago 分析

信任信号

最近提交2 days ago
星标10.5k
许可证Apache-2.0
状态
查看源代码