Zum Hauptinhalt springen
Dieser Inhalt ist noch nicht in Ihrer Sprache verfügbar und wird auf Englisch angezeigt.

Huggingface Local Models

Plugin Verifiziert Aktiv

Use to select models to run locally with llama.cpp and GGUF on CPU, Mac Metal, CUDA, or ROCm. Covers finding GGUFs, quant selection, running servers, exact GGUF file lookup, conversion, and OpenAI-compatible local serving.

Zweck

To enable users to easily run large language models locally on their own hardware, leveraging optimized tools like llama.cpp and Hugging Face's model repository.

Funktionen

  • Select local LLMs with llama.cpp and GGUF
  • Support for CPU, Mac Metal, CUDA, and ROCm
  • Find and select appropriate GGUF models and quantizations
  • Run local LLM servers and CLI interfaces
  • Convert models when GGUF is not directly available

Anwendungsfälle

  • Running LLMs locally for privacy or cost savings.
  • Experimenting with different local LLM configurations and hardware.
  • Developing applications that require a local inference backend.

Nicht-Ziele

  • Providing a managed cloud LLM service.
  • Acting as a general-purpose LLM API wrapper.
  • Abstracting away the underlying llama.cpp or Hugging Face Hub tooling entirely.

Installation

Zuerst Marketplace hinzufügen

/plugin marketplace add huggingface/skills
/plugin install huggingface-local-models@huggingface-skills

Qualitätspunktzahl

Verifiziert
99 /100
Analysiert about 18 hours ago

Vertrauenssignale

Letzter Commit2 days ago
Sterne10.5k
LizenzApache-2.0
Status
Quellcode ansehen