[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"extension-skill-digitalsamba-moviepy-en":3,"guides-for-digitalsamba-moviepy":398,"similar-k174hwztefbn53t86t8rngmeb186mged-en":399},{"_creationTime":4,"_id":5,"children":6,"community":7,"display":9,"evaluation":15,"identity":246,"isFallback":241,"parentExtension":251,"providers":252,"relations":257,"repo":259,"tags":394,"workflow":395},1778686219732.8716,"k174hwztefbn53t86t8rngmeb186mged",[],{"reviewCount":8},0,{"description":10,"installMethods":11,"name":13,"sourceUrl":14},"Python video composition with moviepy 2.x — overlaying deterministic text on AI-generated video (LTX-2, SadTalker), compositing clips, single-file build.py video projects. Use when adding labels/captions/lower-thirds to LTX-2 or SadTalker outputs, building short ad-style spots in pure Python without Remotion, or doing programmatic video composition. Triggers include text overlay on video, label LTX-2 clip, caption SadTalker output, lower third, build.py video, moviepy, Python video composition, sub-30s ad spot.",{"claudeCode":12},"digitalsamba/claude-code-video-toolkit","moviepy","https://github.com/digitalsamba/claude-code-video-toolkit",{"_creationTime":16,"_id":17,"extensionId":5,"locale":18,"result":19,"trustSignals":227,"workflow":244},1778686396933.365,"kn7c9kbgt7783ngzsemj25xdzs86m4p6","en",{"checks":20,"evaluatedAt":196,"extensionSummary":197,"features":198,"nonGoals":204,"promptVersionExtension":209,"promptVersionScoring":210,"purpose":211,"rationale":212,"score":213,"summary":214,"tags":215,"targetMarket":220,"tier":221,"useCases":222},[21,26,29,32,36,39,44,49,52,55,59,63,66,70,73,76,79,82,85,88,92,96,100,104,108,111,114,117,121,124,127,130,133,136,139,143,147,151,154,158,161,164,167,170,174,177,180,183,186,189,193],{"category":22,"check":23,"severity":24,"summary":25},"Practical Utility","Problem relevance","pass","The description clearly states the problem of overlaying deterministic text on AI-generated video from tools like LTX-2 and SadTalker, and the use case for programmatic video composition.",{"category":22,"check":27,"severity":24,"summary":28},"Unique selling proposition","The skill offers significant value over basic prompting by providing a robust moviepy-based workflow for deterministic text overlays and single-file Python video projects, addressing the limitations of AI-generated text and the complexity of tools like Remotion.",{"category":22,"check":30,"severity":24,"summary":31},"Production readiness","The skill is production-ready, offering clear examples, documented dependencies, and a focused workflow for adding deterministic text overlays to AI-generated video.",{"category":33,"check":34,"severity":24,"summary":35},"Scope","Single responsibility principle","The extension focuses on Python video composition using moviepy, specifically for text overlays on AI-generated video and programmatic video projects, aligning with its described scope.",{"category":33,"check":37,"severity":24,"summary":38},"Description quality","The description accurately reflects the skill's capabilities and use cases, detailing the use of moviepy for text overlays on AI-generated video and programmatic video composition, without being misleading.",{"category":40,"check":41,"severity":42,"summary":43},"Invocation","Scoped tools","not_applicable","This is a skill, not a tool-based extension, so the concept of scoped tools does not apply.",{"category":45,"check":46,"severity":47,"summary":48},"Documentation","Configuration & parameter reference","info","While dependencies are listed in requirements.txt, explicit documentation on specific moviepy parameters or configurations beyond the provided examples is not detailed.",{"category":33,"check":50,"severity":42,"summary":51},"Tool naming","This is a skill with a primary function, not a set of distinct tools with names to evaluate.",{"category":33,"check":53,"severity":42,"summary":54},"Minimal I/O surface","As a skill, it does not expose tools with parameters or response shapes in the same way a tool-based extension would.",{"category":56,"check":57,"severity":24,"summary":58},"License","License usability","The extension is licensed under the MIT license, as confirmed by the LICENSE file.",{"category":60,"check":61,"severity":24,"summary":62},"Maintenance","Commit recency","The last commit was on 2026-05-11, which is recent.",{"category":60,"check":64,"severity":24,"summary":65},"Dependency Management","Dependencies are managed via requirements.txt, and the README indicates they can be installed with `python3 -m pip install -r tools/requirements.txt`.",{"category":67,"check":68,"severity":42,"summary":69},"Security","Secret Management","The skill does not handle or expose any secrets.",{"category":67,"check":71,"severity":24,"summary":72},"Injection","The skill focuses on video composition and text rendering using Python libraries, and does not appear to load or execute untrusted third-party data as instructions.",{"category":67,"check":74,"severity":24,"summary":75},"Transitive Supply-Chain Grenades","The skill relies on standard Python libraries and does not fetch external content at runtime to be interpreted as instructions.",{"category":67,"check":77,"severity":24,"summary":78},"Sandbox Isolation","The skill operates on provided video files and generates new ones, without attempting to modify files outside its designated project or bundle scope.",{"category":67,"check":80,"severity":24,"summary":81},"Sandbox escape primitives","No detached-process spawns or deny-retry loops were found in the provided script snippets.",{"category":67,"check":83,"severity":24,"summary":84},"Data Exfiltration","The skill's operations are local file manipulations and do not involve sending user data or credentials to third parties.",{"category":67,"check":86,"severity":24,"summary":87},"Hidden Text Tricks","The bundled content and description are free of hidden-steering tricks, invisible characters, or obfuscated text.",{"category":89,"check":90,"severity":24,"summary":91},"Hooks","Opaque code execution","The provided Python code snippets are clear and readable, with no obfuscation, base64 payloads, or runtime script fetching.",{"category":93,"check":94,"severity":24,"summary":95},"Portability","Structural Assumption","The skill assumes a standard Python environment and uses relative paths for input/output files, as demonstrated in examples.",{"category":97,"check":98,"severity":24,"summary":99},"Trust","Issues Attention","1 issue opened and 4 closed in the last 90 days, with a closure rate of approximately 80%, indicating good maintainer responsiveness.",{"category":101,"check":102,"severity":24,"summary":103},"Versioning","Release Management","The project has a GitHub release tag and indicates a version, aligning with good release management practices.",{"category":105,"check":106,"severity":47,"summary":107},"Code Execution","Validation","While the code uses Pillow and moviepy, explicit validation schemas like Zod or Pydantic for input parameters are not evident in the provided snippets.",{"category":67,"check":109,"severity":24,"summary":110},"Unguarded Destructive Operations","The skill's operations involve creating new video files, which are not considered destructive in the context of file deletion or system modification.",{"category":105,"check":112,"severity":24,"summary":113},"Error Handling","The provided snippets demonstrate basic error handling, such as checking for file existence and clean exits, and the README suggests clear error messages for missing dependencies.",{"category":105,"check":115,"severity":42,"summary":116},"Logging","This skill does not appear to perform destructive actions or outbound calls that would necessitate local audit logging.",{"category":118,"check":119,"severity":42,"summary":120},"Compliance","GDPR","The skill operates on video files and text overlays; it does not process personal data.",{"category":118,"check":122,"severity":24,"summary":123},"Target market","The skill is globally applicable and does not exhibit any regional or jurisdictional limitations.",{"category":93,"check":125,"severity":24,"summary":126},"Runtime stability","The skill assumes a standard Python 3 environment and common libraries, with clear installation instructions, making it portable across POSIX-compliant systems.",{"category":45,"check":128,"severity":24,"summary":129},"README","The README provides a comprehensive overview of the toolkit, its purpose, features, and setup instructions.",{"category":33,"check":131,"severity":42,"summary":132},"Tool surface size","This is a single skill, not a collection of multiple tools to evaluate for surface size.",{"category":40,"check":134,"severity":42,"summary":135},"Overlapping near-synonym tools","As a single skill, there are no multiple tools with overlapping functionalities to evaluate.",{"category":45,"check":137,"severity":24,"summary":138},"Phantom features","All advertised features related to moviepy video composition are supported by the provided code and documentation.",{"category":140,"check":141,"severity":24,"summary":142},"Install","Installation instruction","The README provides clear installation instructions using pip and includes runnable examples for quick testing.",{"category":144,"check":145,"severity":24,"summary":146},"Errors","Actionable error messages","The skill provides actionable error messages, such as guiding users to install missing dependencies or indicating issues with file paths.",{"category":148,"check":149,"severity":24,"summary":150},"Execution","Pinned dependencies","Dependencies are declared in requirements.txt and the README suggests installing them via pip, indicating pinned dependencies.",{"category":33,"check":152,"severity":42,"summary":153},"Dry-run preview","The skill focuses on local file manipulation (video composition) and does not involve state-changing operations or outbound data sending that would require a dry-run mode.",{"category":155,"check":156,"severity":42,"summary":157},"Protocol","Idempotent retry & timeouts","The skill operates locally on files and does not involve remote calls or state-changing operations requiring idempotency or timeouts.",{"category":118,"check":159,"severity":42,"summary":160},"Telemetry opt-in","The skill does not appear to emit any telemetry.",{"category":40,"check":162,"severity":24,"summary":163},"Precise Purpose","The description clearly defines the skill's purpose: Python video composition with moviepy for deterministic text overlays on AI-generated video and single-file projects.",{"category":40,"check":165,"severity":24,"summary":166},"Concise Frontmatter","The frontmatter is concise and effectively summarizes the skill's core capability and usage scenarios.",{"category":45,"check":168,"severity":24,"summary":169},"Concise Body","The SKILL.md body is well-structured and avoids excessive length, with deeper material appropriately deferred to examples or referenced concepts.",{"category":171,"check":172,"severity":24,"summary":173},"Context","Progressive Disclosure","While not strictly using separate reference files for sub-tasks, the SKILL.md is well-organized and examples are provided, offering a good level of progressive disclosure.",{"category":171,"check":175,"severity":42,"summary":176},"Forked exploration","This skill performs a direct video composition task and does not involve deep exploration requiring a forked context.",{"category":22,"check":178,"severity":24,"summary":179},"Usage examples","Runnable examples for text overlay, data visualization, and ad spots are provided in the SKILL.md and README, demonstrating the skill's capabilities.",{"category":22,"check":181,"severity":24,"summary":182},"Edge cases","The skill addresses edge cases such as the TextClip bug and provides workarounds, along with notes on dependency issues and audio handling.",{"category":105,"check":184,"severity":42,"summary":185},"Tool Fallback","This skill does not rely on external MCP servers or tools that would require a fallback mechanism.",{"category":93,"check":187,"severity":24,"summary":188},"Stack assumptions","The skill clearly states its dependency on Python 3 and necessary libraries (moviepy, Pillow, matplotlib) with installation instructions.",{"category":190,"check":191,"severity":24,"summary":192},"Safety","Halt on unexpected state","The skill's installation instructions and example error handling suggest it would halt gracefully on unexpected states, such as missing dependencies.",{"category":93,"check":194,"severity":24,"summary":195},"Cross-skill coupling","The moviepy skill operates independently and clearly references other skills (like LTX-2, SadTalker, ffmpeg) as distinct components in the workflow, rather than relying on them implicitly.",1778686396630,"This skill leverages the moviepy library in Python to overlay deterministic text onto AI-generated video outputs from models like LTX-2 and SadTalker. It supports compositing clips and building single-file Python video projects, providing a reliable method for adding labels, captions, and lower thirds when AI-generated text is insufficient.",[199,200,201,202,203],"Overlay deterministic text on AI-generated video (LTX-2, SadTalker)","Compose video clips programmatically","Build single-file Python video projects","Use PIL for text rendering workaround to avoid moviepy bugs","Implement audio-anchored timelines for precise timing",[205,206,207,208],"Providing a full-featured video editing suite with a GUI","Handling complex, multi-brand design system reuse typical of Remotion","Generating AI video content itself (relies on LTX-2, SadTalker outputs)","Bulk SRT-driven subtitling (prefers ffmpeg for this)","3.0.0","4.4.0","To enable users to reliably add deterministic text overlays, captions, and labels to AI-generated video content, ensuring accuracy and legibility, and to facilitate programmatic video composition in Python.","The extension is well-documented, production-ready, and secure, with clear examples and use cases. The moviepy skill is focused and robust, with excellent handling of edge cases and dependencies.",97,"A robust Python skill for deterministic video composition and text overlays on AI-generated content.",[216,217,13,218,219],"video","python","composition","ai-video","global","verified",[223,224,225,226],"Adding labels, captions, or lower-thirds to LTX-2 or SadTalker video outputs","Building short ad-style video spots in pure Python","Compositing data-driven visuals (e.g., matplotlib animations) with trustworthy text","Ensuring accurate text in news, documentary, or tutorial videos",{"codeQuality":228,"collectedAt":230,"documentation":231,"maintenance":234,"security":240,"testCoverage":243},{"hasLockfile":229},true,1778686374876,{"descriptionLength":232,"readmeSize":233},516,18886,{"closedIssues90d":235,"forks":236,"hasChangelog":229,"openIssues90d":237,"pushedAt":238,"stars":239},4,187,1,1778501732000,1137,{"hasNpmPackage":241,"license":242,"smitheryVerified":241},false,"MIT",{"hasCi":229,"hasTests":241},{"updatedAt":245},1778686396933,{"basePath":247,"githubOwner":248,"githubRepo":249,"locale":18,"slug":13,"type":250},".claude/skills/moviepy","digitalsamba","claude-code-video-toolkit","skill",null,{"evaluate":253,"extract":255},{"promptVersionExtension":209,"promptVersionScoring":210,"score":213,"tags":254,"targetMarket":220,"tier":221},[216,217,13,218,219],{"commitSha":256},"HEAD",{"repoId":258},"kd70r97eght58pp9f1x8scdagd86n32q",{"_creationTime":260,"_id":258,"identity":261,"providers":262,"workflow":390},1778686211924.9185,{"githubOwner":248,"githubRepo":249,"sourceUrl":14},{"classify":263,"discover":373,"github":376},{"commitSha":256,"extensions":264},[265,276,287,296,304,312,317,326,340,349,357,365],{"basePath":266,"description":267,"displayName":268,"installMethods":269,"rationale":270,"selectedPaths":271,"source":275,"sourceLanguage":18,"type":250},".claude/skills/acestep","AI music generation with ACE-Step 1.5 — background music, vocal tracks, covers, stem extraction, audio repainting, and continuation for video production. Use when generating music, soundtracks, jingles, or working with audio stems. Triggers include background music, soundtrack, jingle, music generation, stem extraction, cover, style transfer, repaint, continuation, or musical composition tasks.","acestep",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/acestep/SKILL.md",[272],{"path":273,"priority":274},"SKILL.md","mandatory","rule",{"basePath":277,"description":278,"displayName":279,"installMethods":280,"rationale":281,"selectedPaths":282,"source":275,"sourceLanguage":18,"type":250},".claude/skills/elevenlabs","Generate AI voiceovers, sound effects, and music using ElevenLabs APIs. Use when creating audio content for videos, podcasts, or games. Triggers include generating voiceovers, narration, dialogue, sound effects from descriptions, background music, soundtrack generation, voice cloning, or any audio synthesis task.","elevenlabs",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/elevenlabs/SKILL.md",[283,284],{"path":273,"priority":274},{"path":285,"priority":286},"reference.md","medium",{"basePath":288,"description":289,"displayName":290,"installMethods":291,"rationale":292,"selectedPaths":293,"source":275,"sourceLanguage":18,"type":250},".claude/skills/ffmpeg","Video and audio processing with FFmpeg. Use for format conversion, resizing, compression, audio extraction, and preparing assets for Remotion. Triggers include converting GIF to MP4, resizing video, extracting audio, compressing files, or any media transformation task.","ffmpeg",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/ffmpeg/SKILL.md",[294,295],{"path":273,"priority":274},{"path":285,"priority":286},{"basePath":297,"description":298,"displayName":299,"installMethods":300,"rationale":301,"selectedPaths":302,"source":275,"sourceLanguage":18,"type":250},".claude/skills/frontend-design","Create distinctive, production-grade frontend interfaces with high design quality. Use this skill when the user asks to build web components, pages, or applications. Generates creative, polished code that avoids generic AI aesthetics.","frontend-design",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/frontend-design/SKILL.md",[303],{"path":273,"priority":274},{"basePath":305,"description":306,"displayName":307,"installMethods":308,"rationale":309,"selectedPaths":310,"source":275,"sourceLanguage":18,"type":250},".claude/skills/ltx2","AI video generation with LTX-2.3 22B — text-to-video, image-to-video clips for video production. Use when generating video clips, animating images, creating b-roll, animated backgrounds, or motion content. Triggers include video generation, animate image, b-roll, motion, video clip, text-to-video, image-to-video.","ltx2",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/ltx2/SKILL.md",[311],{"path":273,"priority":274},{"basePath":247,"description":10,"displayName":13,"installMethods":313,"rationale":314,"selectedPaths":315,"source":275,"sourceLanguage":18,"type":250},{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/moviepy/SKILL.md",[316],{"path":273,"priority":274},{"basePath":318,"description":319,"displayName":320,"installMethods":321,"rationale":322,"selectedPaths":323,"source":275,"sourceLanguage":18,"type":250},".claude/skills/playwright-recording","Record browser interactions as video using Playwright. Use for capturing demo videos, app walkthroughs, and UI flows for Remotion videos. Triggers include recording a demo, capturing browser video, screen recording a website, or creating walkthrough footage.","playwright-recording",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/playwright-recording/SKILL.md",[324,325],{"path":273,"priority":274},{"path":285,"priority":286},{"basePath":327,"description":328,"displayName":329,"installMethods":330,"rationale":331,"selectedPaths":332,"source":275,"sourceLanguage":18,"type":250},".claude/skills/qwen-edit","AI image editing prompting patterns for Qwen-Image-Edit. Use when editing photos while preserving identity, reframing cropped images, changing clothing or accessories, adjusting poses, applying style transfers, or character transformations. Provides prompt patterns, parameter tuning, and examples.","qwen-edit",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/qwen-edit/SKILL.md",[333,334,336,338],{"path":273,"priority":274},{"path":335,"priority":286},"examples.md",{"path":337,"priority":286},"parameters.md",{"path":339,"priority":286},"prompting.md",{"basePath":341,"description":342,"displayName":343,"installMethods":344,"rationale":345,"selectedPaths":346,"source":275,"sourceLanguage":18,"type":250},".claude/skills/remotion","Toolkit-specific Remotion patterns — custom transitions, shared components, and project conventions. For core Remotion framework knowledge (hooks, animations, rendering, etc.), see the `remotion-official` skill.","remotion",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/remotion/SKILL.md",[347,348],{"path":273,"priority":274},{"path":285,"priority":286},{"basePath":350,"description":351,"displayName":352,"installMethods":353,"rationale":354,"selectedPaths":355,"source":275,"sourceLanguage":18,"type":250},".claude/skills/remotion-official","Best practices for Remotion - Video creation in React","remotion-best-practices",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/remotion-official/SKILL.md",[356],{"path":273,"priority":274},{"basePath":358,"description":359,"displayName":360,"installMethods":361,"rationale":362,"selectedPaths":363,"source":275,"sourceLanguage":18,"type":250},".claude/skills/runpod","Cloud GPU processing via RunPod serverless. Use when setting up RunPod endpoints, deploying Docker images, managing GPU resources, troubleshooting endpoint issues, or understanding costs. Covers all 5 toolkit images (qwen-edit, realesrgan, propainter, sadtalker, qwen3-tts).","runpod",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/runpod/SKILL.md",[364],{"path":273,"priority":274},{"basePath":366,"description":367,"displayName":368,"installMethods":369,"rationale":370,"selectedPaths":371,"source":275,"sourceLanguage":18,"type":250},"skills/openclaw-video-toolkit","Create professional videos autonomously using claude-code-video-toolkit — AI voiceovers, image generation, music, talking heads, and Remotion rendering.","openclaw-video-toolkit",{"claudeCode":12},"SKILL.md frontmatter at skills/openclaw-video-toolkit/SKILL.md",[372],{"path":273,"priority":274},{"sources":374},[375],"manual",{"closedIssues90d":235,"description":377,"forks":236,"license":242,"openIssues90d":237,"pushedAt":238,"readmeSize":233,"stars":239,"topics":378},"AI-native video production toolkit for Claude Code",[379,380,381,279,382,383,343,384,385,386,387,388,389],"ai-video-generator","claude-code","developer-tools","playwright","programmatic-video","text-to-speech","video-editing","video-production","open-source","qwen-tts","openclaw",{"classifiedAt":391,"discoverAt":392,"extractAt":393,"githubAt":393,"updatedAt":391},1778686219532,1778686211925,1778686217771,[219,218,13,217,216],{"evaluatedAt":245,"extractAt":396,"updatedAt":397},1778686219732,1778686649357,[],[400,429,457,488,516,546],{"_creationTime":401,"_id":402,"community":403,"display":404,"identity":410,"providers":414,"relations":422,"tags":424,"workflow":425},1778695321983.4885,"k17ebhanfep7e674dkqscx6dt586mf11",{"reviewCount":8},{"description":405,"installMethods":406,"name":408,"sourceUrl":409},"Generate short AI videos from text or images — text-to-video, image-to-video, and reference-based generation — with zero API key setup. Use when the user wants to create a video clip, animate an image, or generate video from a description.\n",{"claudeCode":407},"pexoai/pexo-skills","videoagent-video-studio","https://github.com/pexoai/pexo-skills",{"basePath":411,"githubOwner":412,"githubRepo":413,"locale":18,"slug":408,"type":250},"skills/videoagent-video-studio","pexoai","pexo-skills",{"evaluate":415,"extract":421},{"promptVersionExtension":209,"promptVersionScoring":210,"score":416,"tags":417,"targetMarket":220,"tier":221},100,[216,219,418,419,420],"text-to-video","image-to-video","generation",{"commitSha":256},{"repoId":423},"kd7a11fjbvhz4653zn8wdan1bd86mrja",[219,420,419,418,216],{"evaluatedAt":426,"extractAt":427,"updatedAt":428},1778695435906,1778695321983,1778695497057,{"_creationTime":430,"_id":431,"community":432,"display":433,"identity":438,"providers":442,"relations":450,"tags":453,"workflow":454},1778685615701.8425,"k1707ctze9p8fn1e339nd2czjn86m6p5",{"reviewCount":8},{"description":434,"installMethods":435,"name":216,"sourceUrl":437},"When the user wants to create, generate, or produce video content using AI tools or programmatic frameworks. Also use when the user mentions 'video production,' 'AI video,' 'Remotion,' 'Hyperframes,' 'HeyGen,' 'Synthesia,' 'Veo,' 'Runway,' 'Kling,' 'Pika,' 'video generation,' 'AI avatar,' 'talking head video,' 'programmatic video,' 'video template,' 'explainer video,' 'product demo video,' 'video pipeline,' or 'make me a video.' Use this for video creation, generation, and production workflows. For video content strategy and what to post, see social-content. For paid video ad creative, see ad-creative.",{"claudeCode":436},"coreyhaines31/marketingskills","https://github.com/coreyhaines31/marketingskills",{"basePath":439,"githubOwner":440,"githubRepo":441,"locale":18,"slug":216,"type":250},"skills/video","coreyhaines31","marketingskills",{"evaluate":443,"extract":449},{"promptVersionExtension":209,"promptVersionScoring":210,"score":416,"tags":444,"targetMarket":220,"tier":221},[216,445,219,343,446,447,448],"production","heygen","runway","content-creation",{"commitSha":256},{"parentExtensionId":451,"repoId":452},"k175jvka8cxxkf91gk8qy25r8186npjr","kd7a4vjty5ay3s25r82cm72wdn86nmg0",[219,448,446,445,343,447,216],{"evaluatedAt":455,"extractAt":456,"updatedAt":455},1778686582142,1778685615701,{"_creationTime":458,"_id":459,"community":460,"display":461,"identity":467,"providers":471,"relations":481,"tags":484,"workflow":485},1778695548458.3403,"k17874jrpzpfq3ftmmahe017n586nmy2",{"reviewCount":8},{"description":462,"installMethods":463,"name":465,"sourceUrl":466},"Compose or analyze sacred music in Hildegard von Bingen's distinctive modal style. Covers modal selection, melodic contour (wide-range melodies), text-setting (syllabic and melismatic), neumatic notation, and liturgical context for antiphons, sequences, and responsories. Use when composing a new piece in Hildegardian style, analyzing an existing chant for structure and mode, researching medieval modal music, preparing to perform or teach Hildegard's music, or setting Latin sacred texts.\n",{"claudeCode":464},"pjt222/agent-almanac","compose-sacred-music","https://github.com/pjt222/agent-almanac",{"basePath":468,"githubOwner":469,"githubRepo":470,"locale":18,"slug":465,"type":250},"skills/compose-sacred-music","pjt222","agent-almanac",{"evaluate":472,"extract":480},{"promptVersionExtension":209,"promptVersionScoring":210,"score":416,"tags":473,"targetMarket":220,"tier":221},[474,475,476,477,478,218,479],"hildegard","sacred-music","chant","modal","musicology","analysis",{"commitSha":256},{"parentExtensionId":482,"repoId":483},"k170h0janaa9kwn7cfgfz2ykss86mmh9","kd7aryv63z61j39n2td1aeqkvh86mh12",[479,476,218,474,477,478,475],{"evaluatedAt":486,"extractAt":487,"updatedAt":486},1778696688896,1778695548458,{"_creationTime":489,"_id":490,"community":491,"display":492,"identity":498,"providers":502,"relations":509,"tags":512,"workflow":513},1778691104676.0088,"k17ay4s9yakhcp91kg0vsc8nwn86ne5d",{"reviewCount":8},{"description":493,"installMethods":494,"name":496,"sourceUrl":497},"Apply a text watermark to a photo using layer-based image composition for brand protection and copyright.",{"claudeCode":495},"iterationlayer/skills","watermark-an-image","https://github.com/iterationlayer/skills",{"basePath":499,"githubOwner":500,"githubRepo":501,"locale":18,"slug":496,"type":250},"skills/watermark-an-image","iterationlayer","skills",{"evaluate":503,"extract":508},{"promptVersionExtension":209,"promptVersionScoring":210,"score":416,"tags":504,"targetMarket":220,"tier":221},[505,506,507,420,218],"image","watermark","api",{"commitSha":256},{"parentExtensionId":510,"repoId":511},"k1721s0xmp59902ybtpakrrffn86n10s","kd76p4g2qmtrkgx99cnab3683d86n4g8",[507,218,420,505,506],{"evaluatedAt":514,"extractAt":515,"updatedAt":514},1778693993734,1778691104676,{"_creationTime":517,"_id":518,"community":519,"display":520,"identity":526,"providers":530,"relations":539,"tags":542,"workflow":543},1778699018122.8064,"k178yxvt3g9djb8ph907q3tv1186n8ex",{"reviewCount":8},{"description":521,"installMethods":522,"name":524,"sourceUrl":525},"Select and optimize embedding models for semantic search and RAG applications. Use when choosing embedding models, implementing chunking strategies, or optimizing embedding quality for specific domains.",{"claudeCode":523},"wshobson/agents","embedding-strategies","https://github.com/wshobson/agents",{"basePath":527,"githubOwner":528,"githubRepo":529,"locale":18,"slug":524,"type":250},"plugins/llm-application-dev/skills/embedding-strategies","wshobson","agents",{"evaluate":531,"extract":538},{"promptVersionExtension":209,"promptVersionScoring":210,"score":416,"tags":532,"targetMarket":220,"tier":221},[533,534,535,536,537,217],"embeddings","rag","semantic-search","vector-databases","llm-applications",{"commitSha":256},{"parentExtensionId":540,"repoId":541},"k1719fyk9jrke6aq23wbyf8ej586n3af","kd74de64zj0axtg5b8t7eqqe2x86nske",[533,537,217,534,535,536],{"evaluatedAt":544,"extractAt":545,"updatedAt":544},1778701750946,1778699018122,{"_creationTime":547,"_id":548,"community":549,"display":550,"identity":556,"providers":560,"relations":570,"tags":573,"workflow":574},1778699647844.0183,"k174bzyyax9v1t5bm0m98bfqyh86m8v8",{"reviewCount":8},{"description":551,"installMethods":552,"name":554,"sourceUrl":555},"AWS Cloud Development Kit (CDK) expert for building cloud infrastructure with TypeScript/Python. Use when creating CDK stacks, defining CDK constructs, implementing infrastructure as code, or when the user mentions CDK, CloudFormation, IaC, cdk synth, cdk deploy, or wants to define AWS infrastructure programmatically. Covers CDK app structure, construct patterns, stack composition, and deployment workflows.",{"claudeCode":553},"zxkane/aws-skills","aws-cdk-development","https://github.com/zxkane/aws-skills",{"basePath":557,"githubOwner":558,"githubRepo":559,"locale":18,"slug":554,"type":250},"plugins/aws-cdk/skills/aws-cdk-development","zxkane","aws-skills",{"evaluate":561,"extract":569},{"promptVersionExtension":209,"promptVersionScoring":210,"score":416,"tags":562,"targetMarket":220,"tier":221},[563,564,565,217,566,567,568],"aws","cdk","typescript","iac","cloudformation","infrastructure",{"commitSha":256},{"parentExtensionId":571,"repoId":572},"k177paz2fgaa1r1kfhgb2esr1n86my7m","kd7708aervxaq6vqq9tdf93s2586mcqy",[563,564,567,566,568,217,565],{"evaluatedAt":575,"extractAt":576,"updatedAt":577},1778699774404,1778699647844,1778699908774]