[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"extension-skill-digitalsamba-qwen-edit-en":3,"guides-for-digitalsamba-qwen-edit":394,"similar-k175r6z7jxdyenwqxp499wmqzh86mmrj-en":395},{"_creationTime":4,"_id":5,"children":6,"community":7,"display":9,"evaluation":15,"identity":242,"isFallback":237,"parentExtension":247,"providers":248,"relations":253,"repo":255,"tags":390,"workflow":391},1778686219732.872,"k175r6z7jxdyenwqxp499wmqzh86mmrj",[],{"reviewCount":8},0,{"description":10,"installMethods":11,"name":13,"sourceUrl":14},"AI image editing prompting patterns for Qwen-Image-Edit. Use when editing photos while preserving identity, reframing cropped images, changing clothing or accessories, adjusting poses, applying style transfers, or character transformations. Provides prompt patterns, parameter tuning, and examples.",{"claudeCode":12},"digitalsamba/claude-code-video-toolkit","qwen-edit","https://github.com/digitalsamba/claude-code-video-toolkit",{"_creationTime":16,"_id":17,"extensionId":5,"locale":18,"result":19,"trustSignals":223,"workflow":240},1778686435222.7124,"kn7bkx2v1tt479mrgyb7x9c94n86nqg5","en",{"checks":20,"evaluatedAt":192,"extensionSummary":193,"features":194,"nonGoals":200,"promptVersionExtension":204,"promptVersionScoring":205,"purpose":206,"rationale":207,"score":208,"summary":209,"tags":210,"targetMarket":216,"tier":217,"useCases":218},[21,26,29,32,36,39,43,47,50,53,57,61,64,68,71,74,77,80,83,86,90,94,98,102,106,109,113,117,121,124,127,130,133,136,139,143,147,150,153,157,160,163,166,169,173,176,179,182,185,189],{"category":22,"check":23,"severity":24,"summary":25},"Practical Utility","Problem relevance","pass","The description clearly names a concrete user problem related to AI image editing and specifies the type of photos and transformations users can perform.",{"category":22,"check":27,"severity":24,"summary":28},"Unique selling proposition","The skill provides specific prompt patterns, parameter tuning, and examples for Qwen-Image-Edit, going beyond basic LLM capabilities for image manipulation.",{"category":22,"check":30,"severity":24,"summary":31},"Production readiness","The skill is designed to be used in a workflow, with a CLI tool (`tools/image_edit.py`) that interacts with a cloud endpoint, covering the complete lifecycle of prompt-based image editing.",{"category":33,"check":34,"severity":24,"summary":35},"Scope","Single responsibility principle","The extension focuses specifically on AI image editing patterns for Qwen-Image-Edit, a single, coherent domain.",{"category":33,"check":37,"severity":24,"summary":38},"Description quality","The displayed description accurately reflects the skill's capabilities as outlined in the SKILL.md and examples.",{"category":40,"check":41,"severity":24,"summary":42},"Invocation","Scoped tools","The primary tool `tools/image_edit.py` is a specific script for image editing, not a general arbitrary command executor.",{"category":44,"check":45,"severity":24,"summary":46},"Documentation","Configuration & parameter reference","The `parameters.md` file comprehensively documents all available parameters, their ranges, and recommendations.",{"category":33,"check":48,"severity":24,"summary":49},"Tool naming","The main tool `image_edit.py` is descriptively named for its function.",{"category":33,"check":51,"severity":24,"summary":52},"Minimal I/O surface","The `image_edit.py` script takes clearly defined parameters like `--input`, `--prompt`, `--negative`, etc., and does not expose excessive or unnecessary fields.",{"category":54,"check":55,"severity":24,"summary":56},"License","License usability","The extension is licensed under the MIT license, which is permissive and clearly stated in the LICENSE file.",{"category":58,"check":59,"severity":24,"summary":60},"Maintenance","Commit recency","The repository has recent commits, with the latest push on May 11, 2026, indicating active maintenance.",{"category":58,"check":62,"severity":24,"summary":63},"Dependency Management","The project includes a `requirements.txt` and `package.json` (implied by README examples), suggesting dependency management practices are in place.",{"category":65,"check":66,"severity":24,"summary":67},"Security","Secret Management","No secrets are hardcoded or directly exposed in the provided code snippets or documentation.",{"category":65,"check":69,"severity":24,"summary":70},"Injection","The skill's code does not appear to load or execute untrusted third-party data as instructions.",{"category":65,"check":72,"severity":24,"summary":73},"Transitive Supply-Chain Grenades","The skill does not fetch external code or data at runtime that could be manipulated.",{"category":65,"check":75,"severity":24,"summary":76},"Sandbox Isolation","The provided script (`image_edit.py`) operates on input files and does not attempt to modify files outside its designated scope.",{"category":65,"check":78,"severity":24,"summary":79},"Sandbox escape primitives","No evidence of detached process spawns or deny-retry loops is present in the script.",{"category":65,"check":81,"severity":24,"summary":82},"Data Exfiltration","There are no instructions to read or submit confidential data to a third party; outbound calls are documented for AI tools.",{"category":65,"check":84,"severity":24,"summary":85},"Hidden Text Tricks","Bundled content is free of hidden steering tricks, and descriptions use clean, printable ASCII and expected Unicode.",{"category":87,"check":88,"severity":24,"summary":89},"Hooks","Opaque code execution","The `image_edit.py` script is plain Python and not obfuscated or dynamically fetched.",{"category":91,"check":92,"severity":24,"summary":93},"Portability","Structural Assumption","The script operates on specified input files and does not make structural assumptions about the user's project organization.",{"category":95,"check":96,"severity":24,"summary":97},"Trust","Issues Attention","With 1 open and 4 closed issues in the last 90 days, the closure rate is sufficient, indicating good maintainer engagement.",{"category":99,"check":100,"severity":24,"summary":101},"Versioning","Release Management","The repository has GitHub release tags and a CHANGELOG.md, providing clear versioning signals.",{"category":103,"check":104,"severity":24,"summary":105},"Execution","Validation","Input parameters for the `image_edit.py` script are clearly defined and expected to be validated by the underlying library or CLI parsing.",{"category":65,"check":107,"severity":24,"summary":108},"Unguarded Destructive Operations","The image editing operation is not destructive; it produces new output files.",{"category":110,"check":111,"severity":24,"summary":112},"Code Execution","Error Handling","The Python script for image editing is expected to have standard error handling, and the underlying libraries would report issues.",{"category":110,"check":114,"severity":115,"summary":116},"Logging","not_applicable","The skill is primarily focused on image generation and does not perform destructive actions or outbound calls that necessitate a local audit log.",{"category":118,"check":119,"severity":115,"summary":120},"Compliance","GDPR","The skill operates on image files and prompts, not directly on personal data, thus not posing a GDPR risk.",{"category":118,"check":122,"severity":24,"summary":123},"Target market","The skill's functionality is general and does not exhibit any regional or jurisdictional logic, making its target market 'global'.",{"category":91,"check":125,"severity":24,"summary":126},"Runtime stability","The script relies on standard Python libraries and Cloud GPU execution, indicating good cross-platform stability.",{"category":44,"check":128,"severity":24,"summary":129},"README","The README.md file provides a comprehensive overview of the toolkit, including setup, features, and usage.",{"category":33,"check":131,"severity":24,"summary":132},"Tool surface size","The core functionality for this skill is accessed via a single, well-defined script (`image_edit.py`).",{"category":40,"check":134,"severity":24,"summary":135},"Overlapping near-synonym tools","The skill focuses on a single primary tool (`image_edit.py`), avoiding near-synonym redundancy.",{"category":44,"check":137,"severity":24,"summary":138},"Phantom features","All advertised features in the README and SKILL.md for Qwen-Image-Edit have corresponding implementations in the provided scripts and documentation.",{"category":140,"check":141,"severity":24,"summary":142},"Install","Installation instruction","The README provides clear installation instructions and copy-pasteable examples for using the toolkit, including setup for cloud GPUs.",{"category":144,"check":145,"severity":24,"summary":146},"Errors","Actionable error messages","Standard Python error handling is expected, and the CLI tool should provide clear messages for invalid inputs or execution failures.",{"category":103,"check":148,"severity":24,"summary":149},"Pinned dependencies","The project specifies dependencies in `requirements.txt` and `package.json` (implied), allowing for pinned versions.",{"category":33,"check":151,"severity":115,"summary":152},"Dry-run preview","The image editing process generates an output file rather than performing destructive operations, making a dry-run preview unnecessary.",{"category":154,"check":155,"severity":24,"summary":156},"Protocol","Idempotent retry & timeouts","The image generation process is expected to be stateless and have reasonable timeouts handled by the cloud GPU provider or underlying libraries.",{"category":118,"check":158,"severity":24,"summary":159},"Telemetry opt-in","The documentation does not mention any telemetry collection, implying it is either non-existent or strictly opt-in and undocumented.",{"category":40,"check":161,"severity":24,"summary":162},"Precise Purpose","The skill's purpose is precisely defined as AI image editing patterns for Qwen-Image-Edit, with clear use cases and boundaries.",{"category":40,"check":164,"severity":24,"summary":165},"Concise Frontmatter","The frontmatter in SKILL.md is concise and effectively summarizes the core capability and provides trigger phrases.",{"category":44,"check":167,"severity":24,"summary":168},"Concise Body","The SKILL.md is reasonably concise, with deeper material like examples and parameter details delegated to separate files.",{"category":170,"check":171,"severity":24,"summary":172},"Context","Progressive Disclosure","Deeper material such as examples, parameters, and prompting guides are provided in separate Markdown files linked from the main SKILL.md.",{"category":170,"check":174,"severity":115,"summary":175},"Forked exploration","This skill is not an exploration or audit-style skill that would require forked context; it performs a specific image editing task.",{"category":22,"check":177,"severity":24,"summary":178},"Usage examples","The `examples.md` file provides concrete, ready-to-use examples demonstrating various image editing scenarios and their outcomes.",{"category":22,"check":180,"severity":24,"summary":181},"Edge cases","The `examples.md` and `parameters.md` files document failure modes, limitations, and recovery steps for various image editing scenarios.",{"category":110,"check":183,"severity":115,"summary":184},"Tool Fallback","The skill uses standard Python libraries and cloud GPU execution; no external MCP servers are required with undocumented versions.",{"category":186,"check":187,"severity":24,"summary":188},"Safety","Halt on unexpected state","The script is expected to halt and report errors on unexpected states, such as invalid inputs or execution failures.",{"category":91,"check":190,"severity":24,"summary":191},"Cross-skill coupling","This skill is self-contained and does not implicitly rely on other skills; it performs its image editing function independently.",1778686435032,"This skill provides AI image editing capabilities using Qwen-Image-Edit, accessible via a Python CLI tool. It supports various transformations like identity preservation, reframing, style transfer, and character transformations, with detailed documentation on prompting, parameters, and examples.",[195,196,197,198,199],"AI image editing for Qwen-Image-Edit","Identity preservation and character transformations","Reframing cropped images","Style transfer and background replacement guidance","Detailed prompt patterns and parameter tuning",[201,202,203],"Background replacement using single-image methods (prone to artifacts)","Face swapping (cannot reliably preserve identity)","Outpainting (cannot reliably extend canvas)","3.0.0","4.4.0","To enable users to perform advanced AI-driven photo editing tasks with Qwen-Image-Edit, offering specific prompt patterns and parameter tuning for predictable and high-quality results.","All checks passed with a 'pass' or 'not_applicable' severity. The skill is well-documented, has clear use cases, provides examples, and handles parameters and errors effectively.",100,"A high-quality skill for AI image editing with Qwen-Image-Edit, offering detailed prompting patterns and examples.",[211,212,213,214,215],"image-editing","ai","qwen","prompting","python","global","verified",[219,220,221,222],"Editing photos while preserving identity","Reframing cropped images (fixing cut-off heads)","Changing clothing or accessories on subjects","Adjusting poses and applying style transfers",{"codeQuality":224,"collectedAt":226,"documentation":227,"maintenance":230,"security":236,"testCoverage":239},{"hasLockfile":225},true,1778686417227,{"descriptionLength":228,"readmeSize":229},298,18886,{"closedIssues90d":231,"forks":232,"hasChangelog":225,"openIssues90d":233,"pushedAt":234,"stars":235},4,187,1,1778501732000,1137,{"hasNpmPackage":237,"license":238,"smitheryVerified":237},false,"MIT",{"hasCi":225,"hasTests":237},{"updatedAt":241},1778686435222,{"basePath":243,"githubOwner":244,"githubRepo":245,"locale":18,"slug":13,"type":246},".claude/skills/qwen-edit","digitalsamba","claude-code-video-toolkit","skill",null,{"evaluate":249,"extract":251},{"promptVersionExtension":204,"promptVersionScoring":205,"score":208,"tags":250,"targetMarket":216,"tier":217},[211,212,213,214,215],{"commitSha":252},"HEAD",{"repoId":254},"kd70r97eght58pp9f1x8scdagd86n32q",{"_creationTime":256,"_id":254,"identity":257,"providers":258,"workflow":386},1778686211924.9185,{"githubOwner":244,"githubRepo":245,"sourceUrl":14},{"classify":259,"discover":369,"github":372},{"commitSha":252,"extensions":260},[261,272,283,292,300,308,316,325,336,345,353,361],{"basePath":262,"description":263,"displayName":264,"installMethods":265,"rationale":266,"selectedPaths":267,"source":271,"sourceLanguage":18,"type":246},".claude/skills/acestep","AI music generation with ACE-Step 1.5 — background music, vocal tracks, covers, stem extraction, audio repainting, and continuation for video production. Use when generating music, soundtracks, jingles, or working with audio stems. Triggers include background music, soundtrack, jingle, music generation, stem extraction, cover, style transfer, repaint, continuation, or musical composition tasks.","acestep",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/acestep/SKILL.md",[268],{"path":269,"priority":270},"SKILL.md","mandatory","rule",{"basePath":273,"description":274,"displayName":275,"installMethods":276,"rationale":277,"selectedPaths":278,"source":271,"sourceLanguage":18,"type":246},".claude/skills/elevenlabs","Generate AI voiceovers, sound effects, and music using ElevenLabs APIs. Use when creating audio content for videos, podcasts, or games. Triggers include generating voiceovers, narration, dialogue, sound effects from descriptions, background music, soundtrack generation, voice cloning, or any audio synthesis task.","elevenlabs",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/elevenlabs/SKILL.md",[279,280],{"path":269,"priority":270},{"path":281,"priority":282},"reference.md","medium",{"basePath":284,"description":285,"displayName":286,"installMethods":287,"rationale":288,"selectedPaths":289,"source":271,"sourceLanguage":18,"type":246},".claude/skills/ffmpeg","Video and audio processing with FFmpeg. Use for format conversion, resizing, compression, audio extraction, and preparing assets for Remotion. Triggers include converting GIF to MP4, resizing video, extracting audio, compressing files, or any media transformation task.","ffmpeg",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/ffmpeg/SKILL.md",[290,291],{"path":269,"priority":270},{"path":281,"priority":282},{"basePath":293,"description":294,"displayName":295,"installMethods":296,"rationale":297,"selectedPaths":298,"source":271,"sourceLanguage":18,"type":246},".claude/skills/frontend-design","Create distinctive, production-grade frontend interfaces with high design quality. Use this skill when the user asks to build web components, pages, or applications. Generates creative, polished code that avoids generic AI aesthetics.","frontend-design",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/frontend-design/SKILL.md",[299],{"path":269,"priority":270},{"basePath":301,"description":302,"displayName":303,"installMethods":304,"rationale":305,"selectedPaths":306,"source":271,"sourceLanguage":18,"type":246},".claude/skills/ltx2","AI video generation with LTX-2.3 22B — text-to-video, image-to-video clips for video production. Use when generating video clips, animating images, creating b-roll, animated backgrounds, or motion content. Triggers include video generation, animate image, b-roll, motion, video clip, text-to-video, image-to-video.","ltx2",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/ltx2/SKILL.md",[307],{"path":269,"priority":270},{"basePath":309,"description":310,"displayName":311,"installMethods":312,"rationale":313,"selectedPaths":314,"source":271,"sourceLanguage":18,"type":246},".claude/skills/moviepy","Python video composition with moviepy 2.x — overlaying deterministic text on AI-generated video (LTX-2, SadTalker), compositing clips, single-file build.py video projects. Use when adding labels/captions/lower-thirds to LTX-2 or SadTalker outputs, building short ad-style spots in pure Python without Remotion, or doing programmatic video composition. Triggers include text overlay on video, label LTX-2 clip, caption SadTalker output, lower third, build.py video, moviepy, Python video composition, sub-30s ad spot.","moviepy",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/moviepy/SKILL.md",[315],{"path":269,"priority":270},{"basePath":317,"description":318,"displayName":319,"installMethods":320,"rationale":321,"selectedPaths":322,"source":271,"sourceLanguage":18,"type":246},".claude/skills/playwright-recording","Record browser interactions as video using Playwright. Use for capturing demo videos, app walkthroughs, and UI flows for Remotion videos. Triggers include recording a demo, capturing browser video, screen recording a website, or creating walkthrough footage.","playwright-recording",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/playwright-recording/SKILL.md",[323,324],{"path":269,"priority":270},{"path":281,"priority":282},{"basePath":243,"description":10,"displayName":13,"installMethods":326,"rationale":327,"selectedPaths":328,"source":271,"sourceLanguage":18,"type":246},{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/qwen-edit/SKILL.md",[329,330,332,334],{"path":269,"priority":270},{"path":331,"priority":282},"examples.md",{"path":333,"priority":282},"parameters.md",{"path":335,"priority":282},"prompting.md",{"basePath":337,"description":338,"displayName":339,"installMethods":340,"rationale":341,"selectedPaths":342,"source":271,"sourceLanguage":18,"type":246},".claude/skills/remotion","Toolkit-specific Remotion patterns — custom transitions, shared components, and project conventions. For core Remotion framework knowledge (hooks, animations, rendering, etc.), see the `remotion-official` skill.","remotion",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/remotion/SKILL.md",[343,344],{"path":269,"priority":270},{"path":281,"priority":282},{"basePath":346,"description":347,"displayName":348,"installMethods":349,"rationale":350,"selectedPaths":351,"source":271,"sourceLanguage":18,"type":246},".claude/skills/remotion-official","Best practices for Remotion - Video creation in React","remotion-best-practices",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/remotion-official/SKILL.md",[352],{"path":269,"priority":270},{"basePath":354,"description":355,"displayName":356,"installMethods":357,"rationale":358,"selectedPaths":359,"source":271,"sourceLanguage":18,"type":246},".claude/skills/runpod","Cloud GPU processing via RunPod serverless. Use when setting up RunPod endpoints, deploying Docker images, managing GPU resources, troubleshooting endpoint issues, or understanding costs. Covers all 5 toolkit images (qwen-edit, realesrgan, propainter, sadtalker, qwen3-tts).","runpod",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/runpod/SKILL.md",[360],{"path":269,"priority":270},{"basePath":362,"description":363,"displayName":364,"installMethods":365,"rationale":366,"selectedPaths":367,"source":271,"sourceLanguage":18,"type":246},"skills/openclaw-video-toolkit","Create professional videos autonomously using claude-code-video-toolkit — AI voiceovers, image generation, music, talking heads, and Remotion rendering.","openclaw-video-toolkit",{"claudeCode":12},"SKILL.md frontmatter at skills/openclaw-video-toolkit/SKILL.md",[368],{"path":269,"priority":270},{"sources":370},[371],"manual",{"closedIssues90d":231,"description":373,"forks":232,"license":238,"openIssues90d":233,"pushedAt":234,"readmeSize":229,"stars":235,"topics":374},"AI-native video production toolkit for Claude Code",[375,376,377,275,378,379,339,380,381,382,383,384,385],"ai-video-generator","claude-code","developer-tools","playwright","programmatic-video","text-to-speech","video-editing","video-production","open-source","qwen-tts","openclaw",{"classifiedAt":387,"discoverAt":388,"extractAt":389,"githubAt":389,"updatedAt":387},1778686219532,1778686211925,1778686217771,[212,211,214,215,213],{"evaluatedAt":241,"extractAt":392,"updatedAt":393},1778686219732,1778686649691,[],[396,427,454,477,507,536],{"_creationTime":397,"_id":398,"community":399,"display":400,"identity":406,"providers":411,"relations":420,"tags":423,"workflow":424},1778696691708.3274,"k170az7r02e9e2v47mpy80kx6n86nff3",{"reviewCount":8},{"description":401,"installMethods":402,"name":404,"sourceUrl":405},"Detect current market regime using npx neural-trader — bull/bear/ranging/volatile classification with recommended strategy",{"claudeCode":403},"ruvnet/ruflo","Trader Regime","https://github.com/ruvnet/ruflo",{"basePath":407,"githubOwner":408,"githubRepo":409,"locale":18,"slug":410,"type":246},"plugins/ruflo-neural-trader/skills/trader-regime","ruvnet","ruflo","trader-regime",{"evaluate":412,"extract":419},{"promptVersionExtension":204,"promptVersionScoring":205,"score":208,"tags":413,"targetMarket":216,"tier":217},[414,415,416,212,417,418],"finance","trading","market-analysis","typescript","cli",{"commitSha":252,"license":238},{"parentExtensionId":421,"repoId":422},"k17drge8h1fgzchr0p4jaeg33n86mwmy","kd7ed28gj8n0y3msk5dzrp05zs86nqtc",[212,418,414,416,415,417],{"evaluatedAt":425,"extractAt":426,"updatedAt":425},1778701108877,1778696691708,{"_creationTime":428,"_id":429,"community":430,"display":431,"identity":437,"providers":441,"relations":446,"tags":449,"workflow":450},1778693539593.1863,"k173a67a16bpq0e29wjd85v71986nx03",{"reviewCount":8},{"description":432,"installMethods":433,"name":435,"sourceUrl":436},"Domain knowledge for AI trading memory — Outcome-Weighted Memory (OWM) architecture, 5 memory types, recall scoring, and behavioral analysis. Use when recording trades, recalling similar contexts, analyzing performance, or checking behavioral drift. Triggers on \"record trade\", \"remember trade\", \"recall\", \"similar trades\", \"performance\", \"behavioral\", \"disposition\", \"affective state\", \"confidence\".",{"claudeCode":434},"mnemox-ai/tradememory-protocol","trading-memory","https://github.com/mnemox-ai/tradememory-protocol",{"basePath":438,"githubOwner":439,"githubRepo":440,"locale":18,"slug":435,"type":246},"tradememory-plugin/skills/trading-memory","mnemox-ai","tradememory-protocol",{"evaluate":442,"extract":445},{"promptVersionExtension":204,"promptVersionScoring":205,"score":208,"tags":443,"targetMarket":216,"tier":217},[415,212,444,414,215],"memory",{"commitSha":252},{"parentExtensionId":447,"repoId":448},"k170vxkqee48k2xq1v55a025nh86nzn7","kd73z11kfekksxyrs8ds0snacs86ncdy",[212,414,444,215,415],{"evaluatedAt":451,"extractAt":452,"updatedAt":453},1778693719816,1778693539593,1778693833320,{"_creationTime":455,"_id":456,"community":457,"display":458,"identity":462,"providers":465,"relations":472,"tags":473,"workflow":474},1778693539593.1858,"k171p5pgbfbm5g4k5sa3y4cj9s86m6hk",{"reviewCount":8},{"description":459,"installMethods":460,"name":461,"sourceUrl":436},"Domain knowledge for the Evolution Engine — LLM-powered autonomous strategy discovery from raw OHLCV data. Covers the generate-backtest-select-evolve loop, vectorized backtesting, out-of-sample validation, and strategy graduation. Use when discovering trading patterns, running backtests, evolving strategies, or reviewing evolution logs. Triggers on \"evolve\", \"discover patterns\", \"backtest\", \"evolution\", \"strategy generation\", \"candidate strategy\".",{"claudeCode":434},"TradeMemory Protocol",{"basePath":463,"githubOwner":439,"githubRepo":440,"locale":18,"slug":464,"type":246},"tradememory-plugin/skills/evolution-engine","evolution-engine",{"evaluate":466,"extract":471},{"promptVersionExtension":204,"promptVersionScoring":205,"score":208,"tags":467,"targetMarket":216,"tier":217},[415,212,444,468,469,470],"audit","compliance","llm",{"commitSha":252,"license":238},{"parentExtensionId":447,"repoId":448},[212,468,469,470,444,415],{"evaluatedAt":475,"extractAt":452,"updatedAt":476},1778693678813,1778693832942,{"_creationTime":478,"_id":479,"community":480,"display":481,"identity":487,"providers":492,"relations":500,"tags":503,"workflow":504},1778691104676.0042,"k17c25w174y6873nhdh566etts86mv7m",{"reviewCount":8},{"description":482,"installMethods":483,"name":485,"sourceUrl":486},"Transform images with resize, crop, smart crop, upscale, remove background, and 20+ operations.",{"claudeCode":484},"iterationlayer/skills","Image Transformation API","https://github.com/iterationlayer/skills",{"basePath":488,"githubOwner":489,"githubRepo":490,"locale":18,"slug":491,"type":246},"skills/image-transformation-api","iterationlayer","skills","image-transformation-api",{"evaluate":493,"extract":499},{"promptVersionExtension":204,"promptVersionScoring":205,"score":208,"tags":494,"targetMarket":216,"tier":217},[495,496,497,498,212],"image","transformation","editing","api",{"commitSha":252,"license":238},{"parentExtensionId":501,"repoId":502},"k1721s0xmp59902ybtpakrrffn86n10s","kd76p4g2qmtrkgx99cnab3683d86n4g8",[212,498,497,495,496],{"evaluatedAt":505,"extractAt":506,"updatedAt":505},1778693613399,1778691104676,{"_creationTime":508,"_id":509,"community":510,"display":511,"identity":517,"providers":521,"relations":529,"tags":532,"workflow":533},1778693180473.0972,"k1716aj3p8agwq6vwvn5n19v8n86mps9",{"reviewCount":8},{"description":512,"installMethods":513,"name":515,"sourceUrl":516},"Azure AI Document Intelligence SDK for .NET. Extract text, tables, and structured data from documents using prebuilt and custom models. Use for invoice processing, receipt extraction, ID document analysis, and custom document models. Triggers: \"Document Intelligence\", \"DocumentIntelligenceClient\", \"form recognizer\", \"invoice extraction\", \"receipt OCR\", \"document analysis .NET\".\n",{"claudeCode":514},"microsoft/agent-skills","azure-ai-document-intelligence-dotnet","https://github.com/microsoft/agent-skills",{"basePath":518,"githubOwner":519,"githubRepo":520,"locale":18,"slug":515,"type":246},".github/plugins/azure-sdk-dotnet/skills/azure-ai-document-intelligence-dotnet","microsoft","agent-skills",{"evaluate":522,"extract":528},{"promptVersionExtension":204,"promptVersionScoring":205,"score":208,"tags":523,"targetMarket":216,"tier":217},[524,212,525,526,527],"azure","document-intelligence","dotnet","sdk",{"commitSha":252},{"parentExtensionId":530,"repoId":531},"k1795g6t3v2fg9whacs7xkm88186nxr1","kd77czgnv00rfjm815pcc5xx5986n5t8",[212,524,525,526,527],{"evaluatedAt":534,"extractAt":535,"updatedAt":534},1778693591440,1778693180473,{"_creationTime":537,"_id":538,"community":539,"display":540,"identity":544,"providers":547,"relations":557,"tags":559,"workflow":560},1778696691708.3308,"k17d3c35ws96bb55ry97apwm5n86mqp2",{"reviewCount":8},{"description":541,"installMethods":542,"name":543,"sourceUrl":405},"Format prompts for different LLM providers with chat templates and HNSW-powered context retrieval",{"claudeCode":403},"Chat Format",{"basePath":545,"githubOwner":408,"githubRepo":409,"locale":18,"slug":546,"type":246},"plugins/ruflo-ruvllm/skills/chat-format","chat-format",{"evaluate":548,"extract":556},{"promptVersionExtension":204,"promptVersionScoring":205,"score":208,"tags":549,"targetMarket":216,"tier":217},[470,214,550,551,552,553,554,555],"rag","context-retrieval","openai","anthropic","gemini","ollama",{"commitSha":252,"license":238},{"parentExtensionId":558,"repoId":422},"k17ekc0sj70ms9kgkkgr2ypr4s86mz40",[553,551,554,470,555,552,214,550],{"evaluatedAt":561,"extractAt":426,"updatedAt":561},1778701390930]