[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"extension-skill-digitalsamba-ltx2-en":3,"guides-for-digitalsamba-ltx2":402,"similar-k174cmwts8pcx7rj20ryz4xf4n86nqrg-en":403},{"_creationTime":4,"_id":5,"children":6,"community":7,"display":9,"evaluation":15,"identity":251,"isFallback":246,"parentExtension":256,"providers":257,"relations":262,"repo":264,"tags":398,"workflow":399},1778686219732.8713,"k174cmwts8pcx7rj20ryz4xf4n86nqrg",[],{"reviewCount":8},0,{"description":10,"installMethods":11,"name":13,"sourceUrl":14},"AI video generation with LTX-2.3 22B — text-to-video, image-to-video clips for video production. Use when generating video clips, animating images, creating b-roll, animated backgrounds, or motion content. Triggers include video generation, animate image, b-roll, motion, video clip, text-to-video, image-to-video.",{"claudeCode":12},"digitalsamba/claude-code-video-toolkit","ltx2","https://github.com/digitalsamba/claude-code-video-toolkit",{"_creationTime":16,"_id":17,"extensionId":5,"locale":18,"result":19,"trustSignals":232,"workflow":249},1778686374605.4724,"kn73txj3m7g86hnj308y1kjzbd86ma7w","en",{"checks":20,"evaluatedAt":196,"extensionSummary":197,"features":198,"nonGoals":205,"promptVersionExtension":210,"promptVersionScoring":211,"purpose":212,"rationale":213,"score":214,"summary":215,"tags":216,"targetMarket":224,"tier":225,"useCases":226},[21,26,29,32,36,39,43,47,50,53,57,61,64,69,72,75,78,81,84,87,91,95,99,103,107,110,114,117,122,125,128,131,134,137,140,144,148,151,154,158,161,164,167,170,174,177,180,183,186,189,193],{"category":22,"check":23,"severity":24,"summary":25},"Practical Utility","Problem relevance","pass","The description clearly states the problem of AI video generation using LTX-2.3 and identifies specific use cases like text-to-video and image-to-video clips.",{"category":22,"check":27,"severity":24,"summary":28},"Unique selling proposition","The skill leverages LTX-2.3 22B DiT model for AI video generation, offering specialized capabilities beyond standard LLM text generation.",{"category":22,"check":30,"severity":24,"summary":31},"Production readiness","The skill provides a complete workflow for generating video clips, including detailed parameters, style LoRAs, and integration with other tools. Setup instructions are clear.",{"category":33,"check":34,"severity":24,"summary":35},"Scope","Single responsibility principle","The extension focuses on AI video generation using LTX-2.3, with related tools for supporting tasks like image editing and voiceover, maintaining a coherent domain.",{"category":33,"check":37,"severity":24,"summary":38},"Description quality","The displayed description accurately reflects the extension's capabilities, clearly outlining its purpose and use cases for AI video generation.",{"category":40,"check":41,"severity":24,"summary":42},"Invocation","Scoped tools","The extension exposes specific, scoped tools like `ltx2.py` with clearly defined parameters for video generation tasks.",{"category":44,"check":45,"severity":24,"summary":46},"Documentation","Configuration & parameter reference","All parameters for the LTX-2.3 model are clearly documented in the SKILL.md, including defaults and descriptions.",{"category":33,"check":48,"severity":24,"summary":49},"Tool naming","Tool names like `ltx2.py` and `voiceover.py` are descriptive and follow common conventions.",{"category":33,"check":51,"severity":24,"summary":52},"Minimal I/O surface","Input parameters for the `ltx2.py` tool are specific to video generation tasks, and output is a video file.",{"category":54,"check":55,"severity":24,"summary":56},"License","License usability","The extension is licensed under the MIT License, which is a permissive open-source license.",{"category":58,"check":59,"severity":24,"summary":60},"Maintenance","Commit recency","The last commit was on May 11, 2026, which is within the last 3 months.",{"category":58,"check":62,"severity":24,"summary":63},"Dependency Management","The project includes a `requirements.txt` for Python dependencies and `package.json` for Node.js dependencies, indicating management of third-party libraries.",{"category":65,"check":66,"severity":67,"summary":68},"Security","Secret Management","warning","The SKILL.md mentions requiring `MODAL_LTX2_ENDPOINT_URL` in `.env`, and setup instructions mention HuggingFace token and Gemma 3 license acceptance, but there is no explicit guidance on how these secrets are securely handled or logged.",{"category":65,"check":70,"severity":24,"summary":71},"Injection","The skill primarily uses direct tool invocations with structured parameters, reducing the risk of prompt injection through bundled scripts.",{"category":65,"check":73,"severity":24,"summary":74},"Transitive Supply-Chain Grenades","The extension bundles its dependencies and scripts, and cloud GPU calls are managed through specific tool interfaces, avoiding runtime external code execution.",{"category":65,"check":76,"severity":24,"summary":77},"Sandbox Isolation","The provided scripts and tools operate within their defined scopes, focusing on video generation tasks and not attempting to modify files outside their intended directory.",{"category":65,"check":79,"severity":24,"summary":80},"Sandbox escape primitives","No evidence of detached process spawns or deny-retry loops was found in the provided scripts.",{"category":65,"check":82,"severity":24,"summary":83},"Data Exfiltration","The skill's primary function is video generation, and there are no indications of it reading or submitting confidential user data to third parties.",{"category":65,"check":85,"severity":24,"summary":86},"Hidden Text Tricks","The bundled content appears free of hidden steering tricks, and descriptions use standard printable characters.",{"category":88,"check":89,"severity":24,"summary":90},"Hooks","Opaque code execution","The bundled scripts are plain Python and Node.js, with no evidence of obfuscation, base64 payloads, or runtime code fetching.",{"category":92,"check":93,"severity":24,"summary":94},"Portability","Structural Assumption","The setup instructions and script examples assume a standard project structure and command-line usage, providing clear guidance.",{"category":96,"check":97,"severity":24,"summary":98},"Trust","Issues Attention","In the last 90 days, 1 issue was opened and 4 were closed, indicating active maintainer engagement and a healthy closure rate.",{"category":100,"check":101,"severity":24,"summary":102},"Versioning","Release Management","The project has GitHub releases, and the README references versioning information, allowing for clear detection of new versions.",{"category":104,"check":105,"severity":24,"summary":106},"Execution","Validation","The Python scripts validate input parameters like prompts, widths, and heights, ensuring they meet the expected formats and constraints.",{"category":65,"check":108,"severity":24,"summary":109},"Unguarded Destructive Operations","The primary operations involve generating video files, which are not destructive. Configuration steps for cloud GPUs are guided and explicit.",{"category":111,"check":112,"severity":24,"summary":113},"Code Execution","Error Handling","The Python scripts include error handling for parameters and cloud operations, and the documentation outlines limitations and recovery steps.",{"category":111,"check":115,"severity":24,"summary":116},"Logging","The setup instructions and cloud deployment scripts imply logging through Modal's/RunPod's infrastructure, and the `--debug` flag is mentioned for verbose output.",{"category":118,"check":119,"severity":120,"summary":121},"Compliance","GDPR","not_applicable","The extension focuses on video generation and does not appear to operate on personal data.",{"category":118,"check":123,"severity":24,"summary":124},"Target market","The extension is globally applicable, with no specific regional or jurisdictional limitations indicated in the documentation.",{"category":92,"check":126,"severity":24,"summary":127},"Runtime stability","The toolkit specifies Python 3.9+ and Node.js 18+ requirements and provides clear setup instructions for cloud environments, ensuring broad compatibility.",{"category":44,"check":129,"severity":24,"summary":130},"README","The README is comprehensive, providing installation instructions, usage examples, features, and technical details.",{"category":33,"check":132,"severity":24,"summary":133},"Tool surface size","The extension exposes a reasonable number of Python tools and Claude Code commands, all related to video production.",{"category":40,"check":135,"severity":24,"summary":136},"Overlapping near-synonym tools","Tool names and functions are distinct, covering specific aspects of video generation and production without significant overlap.",{"category":44,"check":138,"severity":24,"summary":139},"Phantom features","All advertised features, such as AI video generation, voiceover, and music, have corresponding tools and detailed documentation.",{"category":141,"check":142,"severity":24,"summary":143},"Install","Installation instruction","The README provides clear, step-by-step installation instructions, including setup for cloud GPUs and AI models, with copy-pasteable examples.",{"category":145,"check":146,"severity":24,"summary":147},"Errors","Actionable error messages","The documentation outlines potential failure modes and recovery steps, particularly for cloud GPU setup and tool parameters.",{"category":104,"check":149,"severity":24,"summary":150},"Pinned dependencies","Dependencies are managed via `requirements.txt` and `package.json`, indicating pinned versions for reproducibility.",{"category":33,"check":152,"severity":120,"summary":153},"Dry-run preview","The primary operation is video generation, which is not a state-changing operation in the traditional sense that would benefit from a dry-run flag.",{"category":155,"check":156,"severity":24,"summary":157},"Protocol","Idempotent retry & timeouts","Cloud operations through Modal and RunPod typically handle retries and timeouts, and the script structure suggests discrete, stateless calls for each generation task.",{"category":118,"check":159,"severity":24,"summary":160},"Telemetry opt-in","The extension does not appear to emit telemetry by default; cloud provider integrations and optional local tools would manage their own telemetry policies.",{"category":40,"check":162,"severity":24,"summary":163},"Precise Purpose","The skill's purpose is precisely defined as AI video generation with LTX-2.3, including text-to-video and image-to-video, with clear use cases and triggers.",{"category":40,"check":165,"severity":24,"summary":166},"Concise Frontmatter","The SKILL.md frontmatter is concise, clearly stating the core capability and providing relevant trigger phrases.",{"category":44,"check":168,"severity":24,"summary":169},"Concise Body","The SKILL.md is well-structured and avoids excessive verbosity, delegating deeper material to separate files like `references` and `examples`.",{"category":171,"check":172,"severity":24,"summary":173},"Context","Progressive Disclosure","The documentation uses progressive disclosure effectively, with the main SKILL.md outlining procedures and linking to separate files for detailed information.",{"category":171,"check":175,"severity":120,"summary":176},"Forked exploration","This skill is focused on generation and does not involve deep code review or exploration that would necessitate `context: fork`.",{"category":22,"check":178,"severity":24,"summary":179},"Usage examples","The SKILL.md and README provide numerous clear, ready-to-use examples for text-to-video, image-to-video, and various parameters, demonstrating observable outcomes.",{"category":22,"check":181,"severity":24,"summary":182},"Edge cases","The SKILL.md documentation mentions known limitations like text rendering issues and maximum duration, along with suggestions for recovery (e.g., different seeds, stitching).",{"category":111,"check":184,"severity":24,"summary":185},"Tool Fallback","The extension allows for optional cloud GPU usage with Modal or RunPod, and the `ltx2.py` tool can be run locally if cloud endpoints are not configured, providing a fallback.",{"category":92,"check":187,"severity":24,"summary":188},"Stack assumptions","The project clearly declares requirements for Node.js and Python versions, and specifies cloud environments (Modal/RunPod) for GPU tasks.",{"category":190,"check":191,"severity":24,"summary":192},"Safety","Halt on unexpected state","The setup instructions guide the user through configuration steps, implying that the workflow will halt with clear errors if prerequisites are not met.",{"category":92,"check":194,"severity":24,"summary":195},"Cross-skill coupling","The skill is self-contained and focused on video generation, without implicit reliance on other specific skills being loaded concurrently.",1778686374491,"This skill enables AI video generation using the LTX-2.3 22B DiT model, supporting text-to-video and image-to-video clip creation. It offers detailed parameter controls, style LoRAs, and integrates with cloud GPUs via Modal or RunPod.",[199,200,201,202,203,204],"Text-to-video generation","Image-to-video animation","Customizable video parameters (resolution, duration, quality)","Style LoRA application for aesthetic control","Integration with cloud GPU platforms (Modal, RunPod)","Detailed prompting guide and use case examples",[206,207,208,209],"Generating photorealistic human lip-sync where precision is critical","Producing videos longer than ~8 seconds without stitching","Generating readable text directly within video frames","Providing voiceover or music directly (requires integration with other tools)","3.0.0","4.4.0","To generate AI-powered video clips from text prompts or existing images for various video production needs, such as creating b-roll, animated backgrounds, or motion content.","The extension has a high score due to its comprehensive documentation, clear examples, and well-defined scope. The minor warning regarding secret management handling in .env files prevents a perfect score.",96,"Excellent AI video generation skill with thorough documentation and examples.",[217,218,219,220,221,222,223],"video-generation","ltx-2-3","text-to-video","image-to-video","modal","runpod","ai-video","global","community",[227,228,229,230,231],"Generating atmospheric b-roll clips","Animating still images for motion content","Creating animated slide backgrounds","Generating stylized character cameos","Producing branded intro/outro motion backgrounds",{"codeQuality":233,"collectedAt":235,"documentation":236,"maintenance":239,"security":245,"testCoverage":248},{"hasLockfile":234},true,1778686338096,{"descriptionLength":237,"readmeSize":238},314,18886,{"closedIssues90d":240,"forks":241,"hasChangelog":234,"openIssues90d":242,"pushedAt":243,"stars":244},4,187,1,1778501732000,1137,{"hasNpmPackage":246,"license":247,"smitheryVerified":246},false,"MIT",{"hasCi":234,"hasTests":246},{"updatedAt":250},1778686374605,{"basePath":252,"githubOwner":253,"githubRepo":254,"locale":18,"slug":13,"type":255},".claude/skills/ltx2","digitalsamba","claude-code-video-toolkit","skill",null,{"evaluate":258,"extract":260},{"promptVersionExtension":210,"promptVersionScoring":211,"score":214,"tags":259,"targetMarket":224,"tier":225},[217,218,219,220,221,222,223],{"commitSha":261},"HEAD",{"repoId":263},"kd70r97eght58pp9f1x8scdagd86n32q",{"_creationTime":265,"_id":263,"identity":266,"providers":267,"workflow":394},1778686211924.9185,{"githubOwner":253,"githubRepo":254,"sourceUrl":14},{"classify":268,"discover":377,"github":380},{"commitSha":261,"extensions":269},[270,281,292,301,309,314,322,331,345,354,362,369],{"basePath":271,"description":272,"displayName":273,"installMethods":274,"rationale":275,"selectedPaths":276,"source":280,"sourceLanguage":18,"type":255},".claude/skills/acestep","AI music generation with ACE-Step 1.5 — background music, vocal tracks, covers, stem extraction, audio repainting, and continuation for video production. Use when generating music, soundtracks, jingles, or working with audio stems. Triggers include background music, soundtrack, jingle, music generation, stem extraction, cover, style transfer, repaint, continuation, or musical composition tasks.","acestep",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/acestep/SKILL.md",[277],{"path":278,"priority":279},"SKILL.md","mandatory","rule",{"basePath":282,"description":283,"displayName":284,"installMethods":285,"rationale":286,"selectedPaths":287,"source":280,"sourceLanguage":18,"type":255},".claude/skills/elevenlabs","Generate AI voiceovers, sound effects, and music using ElevenLabs APIs. Use when creating audio content for videos, podcasts, or games. Triggers include generating voiceovers, narration, dialogue, sound effects from descriptions, background music, soundtrack generation, voice cloning, or any audio synthesis task.","elevenlabs",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/elevenlabs/SKILL.md",[288,289],{"path":278,"priority":279},{"path":290,"priority":291},"reference.md","medium",{"basePath":293,"description":294,"displayName":295,"installMethods":296,"rationale":297,"selectedPaths":298,"source":280,"sourceLanguage":18,"type":255},".claude/skills/ffmpeg","Video and audio processing with FFmpeg. Use for format conversion, resizing, compression, audio extraction, and preparing assets for Remotion. Triggers include converting GIF to MP4, resizing video, extracting audio, compressing files, or any media transformation task.","ffmpeg",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/ffmpeg/SKILL.md",[299,300],{"path":278,"priority":279},{"path":290,"priority":291},{"basePath":302,"description":303,"displayName":304,"installMethods":305,"rationale":306,"selectedPaths":307,"source":280,"sourceLanguage":18,"type":255},".claude/skills/frontend-design","Create distinctive, production-grade frontend interfaces with high design quality. Use this skill when the user asks to build web components, pages, or applications. Generates creative, polished code that avoids generic AI aesthetics.","frontend-design",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/frontend-design/SKILL.md",[308],{"path":278,"priority":279},{"basePath":252,"description":10,"displayName":13,"installMethods":310,"rationale":311,"selectedPaths":312,"source":280,"sourceLanguage":18,"type":255},{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/ltx2/SKILL.md",[313],{"path":278,"priority":279},{"basePath":315,"description":316,"displayName":317,"installMethods":318,"rationale":319,"selectedPaths":320,"source":280,"sourceLanguage":18,"type":255},".claude/skills/moviepy","Python video composition with moviepy 2.x — overlaying deterministic text on AI-generated video (LTX-2, SadTalker), compositing clips, single-file build.py video projects. Use when adding labels/captions/lower-thirds to LTX-2 or SadTalker outputs, building short ad-style spots in pure Python without Remotion, or doing programmatic video composition. Triggers include text overlay on video, label LTX-2 clip, caption SadTalker output, lower third, build.py video, moviepy, Python video composition, sub-30s ad spot.","moviepy",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/moviepy/SKILL.md",[321],{"path":278,"priority":279},{"basePath":323,"description":324,"displayName":325,"installMethods":326,"rationale":327,"selectedPaths":328,"source":280,"sourceLanguage":18,"type":255},".claude/skills/playwright-recording","Record browser interactions as video using Playwright. Use for capturing demo videos, app walkthroughs, and UI flows for Remotion videos. Triggers include recording a demo, capturing browser video, screen recording a website, or creating walkthrough footage.","playwright-recording",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/playwright-recording/SKILL.md",[329,330],{"path":278,"priority":279},{"path":290,"priority":291},{"basePath":332,"description":333,"displayName":334,"installMethods":335,"rationale":336,"selectedPaths":337,"source":280,"sourceLanguage":18,"type":255},".claude/skills/qwen-edit","AI image editing prompting patterns for Qwen-Image-Edit. Use when editing photos while preserving identity, reframing cropped images, changing clothing or accessories, adjusting poses, applying style transfers, or character transformations. Provides prompt patterns, parameter tuning, and examples.","qwen-edit",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/qwen-edit/SKILL.md",[338,339,341,343],{"path":278,"priority":279},{"path":340,"priority":291},"examples.md",{"path":342,"priority":291},"parameters.md",{"path":344,"priority":291},"prompting.md",{"basePath":346,"description":347,"displayName":348,"installMethods":349,"rationale":350,"selectedPaths":351,"source":280,"sourceLanguage":18,"type":255},".claude/skills/remotion","Toolkit-specific Remotion patterns — custom transitions, shared components, and project conventions. For core Remotion framework knowledge (hooks, animations, rendering, etc.), see the `remotion-official` skill.","remotion",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/remotion/SKILL.md",[352,353],{"path":278,"priority":279},{"path":290,"priority":291},{"basePath":355,"description":356,"displayName":357,"installMethods":358,"rationale":359,"selectedPaths":360,"source":280,"sourceLanguage":18,"type":255},".claude/skills/remotion-official","Best practices for Remotion - Video creation in React","remotion-best-practices",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/remotion-official/SKILL.md",[361],{"path":278,"priority":279},{"basePath":363,"description":364,"displayName":222,"installMethods":365,"rationale":366,"selectedPaths":367,"source":280,"sourceLanguage":18,"type":255},".claude/skills/runpod","Cloud GPU processing via RunPod serverless. Use when setting up RunPod endpoints, deploying Docker images, managing GPU resources, troubleshooting endpoint issues, or understanding costs. Covers all 5 toolkit images (qwen-edit, realesrgan, propainter, sadtalker, qwen3-tts).",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/runpod/SKILL.md",[368],{"path":278,"priority":279},{"basePath":370,"description":371,"displayName":372,"installMethods":373,"rationale":374,"selectedPaths":375,"source":280,"sourceLanguage":18,"type":255},"skills/openclaw-video-toolkit","Create professional videos autonomously using claude-code-video-toolkit — AI voiceovers, image generation, music, talking heads, and Remotion rendering.","openclaw-video-toolkit",{"claudeCode":12},"SKILL.md frontmatter at skills/openclaw-video-toolkit/SKILL.md",[376],{"path":278,"priority":279},{"sources":378},[379],"manual",{"closedIssues90d":240,"description":381,"forks":241,"license":247,"openIssues90d":242,"pushedAt":243,"readmeSize":238,"stars":244,"topics":382},"AI-native video production toolkit for Claude Code",[383,384,385,284,386,387,348,388,389,390,391,392,393],"ai-video-generator","claude-code","developer-tools","playwright","programmatic-video","text-to-speech","video-editing","video-production","open-source","qwen-tts","openclaw",{"classifiedAt":395,"discoverAt":396,"extractAt":397,"githubAt":397,"updatedAt":395},1778686219532,1778686211925,1778686217771,[223,220,218,221,222,219,217],{"evaluatedAt":250,"extractAt":400,"updatedAt":401},1778686219732,1778686649153,[],[404,433,454,484,512,543],{"_creationTime":405,"_id":406,"community":407,"display":408,"identity":414,"providers":418,"relations":426,"tags":428,"workflow":429},1778695321983.4885,"k17ebhanfep7e674dkqscx6dt586mf11",{"reviewCount":8},{"description":409,"installMethods":410,"name":412,"sourceUrl":413},"Generate short AI videos from text or images — text-to-video, image-to-video, and reference-based generation — with zero API key setup. Use when the user wants to create a video clip, animate an image, or generate video from a description.\n",{"claudeCode":411},"pexoai/pexo-skills","videoagent-video-studio","https://github.com/pexoai/pexo-skills",{"basePath":415,"githubOwner":416,"githubRepo":417,"locale":18,"slug":412,"type":255},"skills/videoagent-video-studio","pexoai","pexo-skills",{"evaluate":419,"extract":425},{"promptVersionExtension":210,"promptVersionScoring":211,"score":420,"tags":421,"targetMarket":224,"tier":424},100,[422,223,219,220,423],"video","generation","verified",{"commitSha":261},{"repoId":427},"kd7a11fjbvhz4653zn8wdan1bd86mrja",[223,423,220,219,422],{"evaluatedAt":430,"extractAt":431,"updatedAt":432},1778695435906,1778695321983,1778695497057,{"_creationTime":434,"_id":435,"community":436,"display":437,"identity":439,"providers":440,"relations":449,"tags":450,"workflow":451},1778686219732.873,"k17crq1jtpdq6v8et37enw77pd86mtmh",{"reviewCount":8},{"description":371,"installMethods":438,"name":372,"sourceUrl":14},{"claudeCode":12},{"basePath":370,"githubOwner":253,"githubRepo":254,"locale":18,"slug":372,"type":255},{"evaluate":441,"extract":448},{"promptVersionExtension":210,"promptVersionScoring":211,"score":442,"tags":443,"targetMarket":224,"tier":424},99,[217,444,348,221,445,446,447],"ai-tools","python","typescript","automation",{"commitSha":261},{"repoId":263},[444,447,221,445,348,446,217],{"evaluatedAt":452,"extractAt":400,"updatedAt":453},1778686517002,1778686650397,{"_creationTime":455,"_id":456,"community":457,"display":458,"identity":464,"providers":468,"relations":477,"tags":479,"workflow":480},1778697542749.5203,"k1778wrb4sjw606cz318m6fc9d86mg0e",{"reviewCount":8},{"description":459,"installMethods":460,"name":462,"sourceUrl":463},"Draft and refine prompts for video generation models (text-to-video and image-to-video), and create character-sheet prompts for image models when the goal is character consistency before image-to-video. Use when a user asks for a \"video prompt\", a model-specific prompt such as Seedance 2.0, Ovi, Sora, Veo 3, Wan 2.2, LTX-2, or LTX-2.3, or a consistent-character prompt such as \"character sheet prompt\", \"character turnaround\", \"character reference sheet\", or \"photographic identity sheet\".",{"claudeCode":461},"Square-Zero-Labs/video-prompting-skill","Video Prompting Skill","https://github.com/Square-Zero-Labs/video-prompting-skill",{"basePath":465,"githubOwner":466,"githubRepo":467,"locale":18,"slug":465,"type":255},"video-prompting","Square-Zero-Labs","video-prompting-skill",{"evaluate":469,"extract":475},{"promptVersionExtension":210,"promptVersionScoring":211,"score":470,"tags":471,"targetMarket":224,"tier":424},95,[217,472,473,474,219,220],"prompt-engineering","character-consistency","ai-art",{"commitSha":261,"license":476},"Apache-2.0",{"repoId":478},"kd7a3nw9bpm9b0jdnkskwdnrvx86ndrj",[474,473,220,472,219,217],{"evaluatedAt":481,"extractAt":482,"updatedAt":483},1778697559479,1778697542749,1778697580777,{"_creationTime":485,"_id":486,"community":487,"display":488,"identity":493,"providers":497,"relations":505,"tags":508,"workflow":509},1778685615701.8425,"k1707ctze9p8fn1e339nd2czjn86m6p5",{"reviewCount":8},{"description":489,"installMethods":490,"name":422,"sourceUrl":492},"When the user wants to create, generate, or produce video content using AI tools or programmatic frameworks. Also use when the user mentions 'video production,' 'AI video,' 'Remotion,' 'Hyperframes,' 'HeyGen,' 'Synthesia,' 'Veo,' 'Runway,' 'Kling,' 'Pika,' 'video generation,' 'AI avatar,' 'talking head video,' 'programmatic video,' 'video template,' 'explainer video,' 'product demo video,' 'video pipeline,' or 'make me a video.' Use this for video creation, generation, and production workflows. For video content strategy and what to post, see social-content. For paid video ad creative, see ad-creative.",{"claudeCode":491},"coreyhaines31/marketingskills","https://github.com/coreyhaines31/marketingskills",{"basePath":494,"githubOwner":495,"githubRepo":496,"locale":18,"slug":422,"type":255},"skills/video","coreyhaines31","marketingskills",{"evaluate":498,"extract":504},{"promptVersionExtension":210,"promptVersionScoring":211,"score":420,"tags":499,"targetMarket":224,"tier":424},[422,500,223,348,501,502,503],"production","heygen","runway","content-creation",{"commitSha":261},{"parentExtensionId":506,"repoId":507},"k175jvka8cxxkf91gk8qy25r8186npjr","kd7a4vjty5ay3s25r82cm72wdn86nmg0",[223,503,501,500,348,502,422],{"evaluatedAt":510,"extractAt":511,"updatedAt":510},1778686582142,1778685615701,{"_creationTime":513,"_id":514,"community":515,"display":516,"identity":522,"providers":526,"relations":536,"tags":539,"workflow":540},1778695548458.3403,"k17874jrpzpfq3ftmmahe017n586nmy2",{"reviewCount":8},{"description":517,"installMethods":518,"name":520,"sourceUrl":521},"Compose or analyze sacred music in Hildegard von Bingen's distinctive modal style. Covers modal selection, melodic contour (wide-range melodies), text-setting (syllabic and melismatic), neumatic notation, and liturgical context for antiphons, sequences, and responsories. Use when composing a new piece in Hildegardian style, analyzing an existing chant for structure and mode, researching medieval modal music, preparing to perform or teach Hildegard's music, or setting Latin sacred texts.\n",{"claudeCode":519},"pjt222/agent-almanac","compose-sacred-music","https://github.com/pjt222/agent-almanac",{"basePath":523,"githubOwner":524,"githubRepo":525,"locale":18,"slug":520,"type":255},"skills/compose-sacred-music","pjt222","agent-almanac",{"evaluate":527,"extract":535},{"promptVersionExtension":210,"promptVersionScoring":211,"score":420,"tags":528,"targetMarket":224,"tier":424},[529,530,531,221,532,533,534],"hildegard","sacred-music","chant","musicology","composition","analysis",{"commitSha":261},{"parentExtensionId":537,"repoId":538},"k170h0janaa9kwn7cfgfz2ykss86mmh9","kd7aryv63z61j39n2td1aeqkvh86mh12",[534,531,533,529,221,532,530],{"evaluatedAt":541,"extractAt":542,"updatedAt":541},1778696688896,1778695548458,{"_creationTime":544,"_id":545,"community":546,"display":547,"identity":552,"providers":555,"relations":562,"tags":564,"workflow":565},1778691316210.005,"k175txezbab4wcy00m9gcdq8qx86mz9k",{"reviewCount":8},{"description":548,"installMethods":549,"name":393,"sourceUrl":551},"Generate images and videos from text with multi-provider routing — supports GPT Image 2.0 (near-perfect text rendering), Nanobanana 2, Seedream 5.0, Midjourney V8.1 (unified photorealistic + anime), Flux 2 Klein (cheap drafts), Seedance 2.0 / Happyhorse 1.0 / Veo 3.1 video, and local ComfyUI workflows. Includes 1,446 curated prompts and style-aware prompt enhancement. Use when users want to create images/videos, design assets, animate photos, enhance prompts, or manage AI art workflows. NOT for: generic chat, code generation, document writing, video editing of existing footage, audio/TTS, or any task unrelated to AI image/video creation.",{"claudeCode":550},"jau123/MeiGen-AI-Design-MCP","https://github.com/jau123/MeiGen-AI-Design-MCP",{"basePath":393,"githubOwner":553,"githubRepo":554,"locale":18,"slug":393,"type":255},"jau123","MeiGen-AI-Design-MCP",{"evaluate":556,"extract":561},{"promptVersionExtension":210,"promptVersionScoring":211,"score":420,"tags":557,"targetMarket":224,"tier":424},[558,217,474,559,472,560],"image-generation","comfyui","design-assistant",{"commitSha":261},{"repoId":563},"kd7b1trrphkh2efj71w89w7qzh86nh22",[474,559,560,558,472,217],{"evaluatedAt":566,"extractAt":567,"updatedAt":568},1778691399174,1778691316210,1778691569625]