Deploy Edge Ai Model
Skill Verifiziert AktivDeploy machine learning models to edge devices using Google AI Edge Gallery, TensorFlow Lite, ONNX Runtime, and MediaPipe. Covers model quantization (INT8/INT4), on-device inference with Gemma 4 models, Android/iOS deployment via AI Edge Gallery, hardware delegate selection (GPU/NPU/DSP), and performance benchmarking on constrained devices. Use when deploying models to mobile phones, IoT devices, or embedded systems where cloud inference is impractical due to latency, cost, or connectivity constraints.
To enable developers to efficiently deploy and optimize machine learning models on resource-constrained edge devices, overcoming limitations of cloud inference.
Funktionen
- Model quantization (INT8/INT4)
- On-device LLM inference (Gemma 4)
- Android/iOS deployment via AI Edge Gallery
- Hardware delegate selection (GPU/NPU/DSP)
- Performance benchmarking on constrained devices
Anwendungsfälle
- Deploying LLMs to mobile phones
- Running AI on IoT devices
- Optimizing inference for embedded systems
- Benchmarking model performance on target hardware
Nicht-Ziele
- Cloud-based model serving
- General machine learning model training
- Building complex UI frameworks for edge apps
Installation
/plugin install agent-almanac@pjt222-agent-almanacQualitätspunktzahl
VerifiziertVertrauenssignale
Ähnliche Erweiterungen
Containerize MCP Server
100Containerize an R-based MCP (Model Context Protocol) server using Docker. Covers mcptools integration, port exposure, stdio vs HTTP transport, and connecting Claude Code to the containerized server. Use when deploying an R MCP server without requiring a local R installation, creating a reproducible MCP server environment, running MCP servers alongside other containerized services, or distributing an MCP server to other developers.
Azure Deploy
100Execute Azure deployments for ALREADY-PREPARED applications that have existing .azure/deployment-plan.md and infrastructure files. DO NOT use this skill when the user asks to CREATE a new application — use azure-prepare instead. This skill runs azd up, azd deploy, terraform apply, and az deployment commands with built-in error recovery. Requires .azure/deployment-plan.md from azure-prepare and validated status from azure-validate. WHEN: "run azd up", "run azd deploy", "execute deployment", "push to production", "push to cloud", "go live", "ship it", "bicep deploy", "terraform apply", "publish to Azure", "launch on Azure". DO NOT USE WHEN: "create and deploy", "build and deploy", "create a new app", "set up infrastructure", "create and deploy to Azure using Terraform" — use azure-prepare for these.
Wrangler
100Cloudflare Workers CLI zum Bereitstellen, Entwickeln und Verwalten von Workers, KV, R2, D1, Vectorize, Hyperdrive, Workers AI, Containern, Queues, Workflows, Pipelines und Secrets Store. Laden Sie dies, bevor Sie `wrangler`-Befehle ausführen, um die korrekte Syntax und die besten Vorgehensweisen sicherzustellen. Bevorzugt die Abfrage von Cloudflare-Dokumenten gegenüber vortrainiertem Wissen.
Devops
100Deploy to Cloudflare (Workers, R2, D1), Docker, GCP (Cloud Run, GKE), Kubernetes (kubectl, Helm). Use for serverless, containers, CI/CD, GitOps, security audit.
Ship Gate
100Pre-production audit that scans a codebase for security, database, deployment, code quality, AI/LLM, dependency, frontend, and observability issues. Intercepts deploy commands and blocks until critical items pass. Stack-agnostic. Use for "run ship gate", "am I ready to ship", "pre-launch audit", "can I deploy", "push to production", "go live checklist", "preflight check". Not for CI/CD setup or infra provisioning.
TimesFM Forecasting
100Zero-shot time series forecasting with Google's TimesFM foundation model. Use for any univariate time series (sales, sensors, energy, vitals, weather) without training a custom model. Supports CSV/DataFrame/array inputs with point forecasts and prediction intervals. Includes a preflight system checker script to verify RAM/GPU before first use.