Cost Benchmark
Skill Verified ActiveRun the corpus benchmark — booster locally, optional Gemini/Sonnet/Opus baselines — and persist a verifiable measured-vs-claimed table
To provide a verifiable, local benchmark for LLM cost performance, ensuring accuracy in claimed costs and facilitating performance audits.
Features
- Run local corpus benchmark
- Support optional Gemini/Sonnet/Opus baselines
- Persist verifiable measured-vs-claimed table
- Documented environment overrides for customization
- Output historical and latest benchmark runs
Use Cases
- Verify release cost claims before publishing
- Confirm new benchmark cases route correctly
- Audit 'claimed upstream' tags by verifying benchmark support
- Compare costs of different LLM models for specific tasks
Non-Goals
- Modifying benchmark corpus data
- Running benchmarks against live production systems
- Automated remediation of cost regressions
Installation
First, add the marketplace
/plugin marketplace add ruvnet/ruflo/plugin install ruflo-cost-tracker@rufloQuality Score
VerifiedTrust Signals
Similar Extensions
Janitor Tokens
100Show how many context window tokens each skill consumes. Use when the user asks about token cost, context budget, skill size, or wants to know which skills waste the most context space.
Cloud Architect
100Designs cloud architectures, creates migration plans, generates cost optimization recommendations, and produces disaster recovery strategies across AWS, Azure, and GCP. Use when designing cloud architectures, planning migrations, or optimizing multi-cloud deployments. Invoke for Well-Architected Framework, cost optimization, disaster recovery, landing zones, security architecture, serverless design.
Chat Format
100Format prompts for different LLM providers with chat templates and HNSW-powered context retrieval
Oh My Claudecode
100Process-first advisor routing for Claude, Codex, or Gemini via `omc ask`, with artifact capture and no raw CLI assembly
Wrap Up Ritual
100End-of-session ritual that audits changes, runs quality checks, captures learnings, and produces a session summary. Use when saying "wrap up", "done for the day", "finish coding", or ending a coding session.
Project Development
100This skill should be used when the user asks to "start an LLM project", "design batch pipeline", "evaluate task-model fit", "structure agent project", or mentions pipeline architecture, agent-assisted development, cost estimation, or choosing between LLM and traditional approaches.