Modal Serverless Gpu
Skill AktivServerless GPU cloud platform for running ML workloads. Use when you need on-demand GPU access without infrastructure management, deploying ML models as APIs, or running batch jobs with automatic scaling.
To enable users to run GPU-intensive ML workloads on-demand without managing infrastructure, by leveraging Modal's serverless platform for deployment and batch processing.
Funktionen
- Serverless GPU access (T4, L4, A10G, A100, H100, etc.)
- On-demand ML model deployment as APIs
- Automatic scaling for batch jobs and inference
- Python-native infrastructure definition
- Sub-second cold starts and container caching
Anwendungsfälle
- Running GPU-intensive ML workloads without managing infrastructure
- Deploying ML models as auto-scaling APIs
- Running batch processing jobs (training, inference, data processing)
- Prototyping ML applications quickly
Nicht-Ziele
- Providing reserved GPU instances
- Orchestrating multi-cloud deployments
- Managing complex multi-service architectures directly
Trust
- warning:Issues AttentionIn the last 90 days, 17 issues were opened and 4 were closed, indicating a closure rate below 50% and a significant number of open issues, suggesting potential delays in maintainer response.
Installation
npx skills add davila7/claude-code-templatesFührt das Vercel skills CLI (skills.sh) via npx aus — benötigt Node.js lokal und mindestens einen installierten skills-kompatiblen Agent (Claude Code, Cursor, Codex, …). Setzt voraus, dass das Repo dem agentskills.io-Format folgt.
Qualitätspunktzahl
Vertrauenssignale
Ähnliche Erweiterungen
Cloudflare Deploy
99Deploy applications and infrastructure to Cloudflare using Workers, Pages, and related platform services. Use when the user asks to deploy, host, publish, or set up a project on Cloudflare.
Render Deploy
99Deploy applications to Render by analyzing codebases, generating render.yaml Blueprints, and providing Dashboard deeplinks. Use when the user wants to deploy, host, publish, or set up their application on Render's cloud platform.
Cost Optimization
98Optimize cloud costs across AWS, Azure, GCP, and OCI through resource rightsizing, tagging strategies, reserved instances, and spending analysis. Use when reducing cloud expenses, analyzing infrastructure costs, or implementing cost governance policies.
Skypilot Multi Cloud Orchestration
98Multi-cloud orchestration for ML workloads with automatic cost optimization. Use when you need to run training or batch jobs across multiple clouds, leverage spot instances with auto-recovery, or optimize GPU costs across providers.
RunPod Cloud GPU
98Cloud-GPU-Verarbeitung über RunPod Serverless. Verwenden Sie dies beim Einrichten von RunPod-Endpunkten, beim Bereitstellen von Docker-Images, beim Verwalten von GPU-Ressourcen, beim Beheben von Endpunktproblemen oder beim Verstehen von Kosten. Beinhaltet alle 5 Toolkit-Images (qwen-edit, realesrgan, propainter, sadtalker, qwen3-tts).
Alterlab Modal
98Part of the AlterLab Academic Skills suite. Run Python code in the cloud with serverless containers, GPUs, and autoscaling. Use when deploying ML models, running batch processing jobs, scheduling compute-intensive tasks, or serving APIs that require GPU acceleration or dynamic scaling.