[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"extension-skill-digitalsamba-ltx2-zh-CN":3,"guides-for-digitalsamba-ltx2":405,"similar-k171jbhaba8hjw4m5951btc5dh86mz1m-zh-CN":406},{"_creationTime":4,"_id":5,"children":6,"community":7,"display":9,"evaluation":15,"identity":250,"isFallback":245,"parentExtension":255,"providers":256,"relations":262,"repo":265,"tags":401,"workflow":402},1778686575608.8682,"k171jbhaba8hjw4m5951btc5dh86mz1m",[],{"reviewCount":8},0,{"description":10,"installMethods":11,"name":13,"sourceUrl":14},"使用 LTX-2.3 22B 进行 AI 视频生成 — 视频制作的文本到视频、图像到视频片段。在生成视频片段、动画图像、制作备用镜头、动画背景或动态内容时使用。触发器包括视频生成、动画图像、备用镜头、动态、视频片段、文本到视频、图像到视频。",{"claudeCode":12},"digitalsamba/claude-code-video-toolkit","ltx2","https://github.com/digitalsamba/claude-code-video-toolkit",{"_creationTime":16,"_id":17,"extensionId":5,"locale":18,"result":19,"trustSignals":231,"workflow":248},1778686575608.8684,"kn7aagakr1bbfskxn78c7p3xr186mqz7","zh-CN",{"checks":20,"evaluatedAt":196,"extensionSummary":197,"features":198,"nonGoals":205,"promptVersionExtension":210,"promptVersionScoring":211,"purpose":212,"rationale":213,"score":214,"summary":215,"tags":216,"tier":224,"useCases":225},[21,26,29,32,36,39,43,47,50,53,57,61,64,69,72,75,78,81,84,87,91,95,99,103,107,110,114,117,122,125,128,131,134,137,140,144,148,151,154,158,161,164,167,170,174,177,180,183,186,189,193],{"category":22,"check":23,"severity":24,"summary":25},"Practical Utility","Problem relevance","pass","描述清楚地说明了使用 LTX-2.3 进行 AI 视频生成的问题，并确定了文本到视频和图像到视频片段等具体用例。",{"category":22,"check":27,"severity":24,"summary":28},"Unique selling proposition","该技能利用 LTX-2.3 22B DiT 模型进行 AI 视频生成，提供了超越标准 LLM 文本生成的专门功能。",{"category":22,"check":30,"severity":24,"summary":31},"Production readiness","该技能提供了完整的视频片段生成工作流程，包括详细参数、样式 LoRA 以及与其他工具的集成。设置说明清晰。",{"category":33,"check":34,"severity":24,"summary":35},"Scope","Single responsibility principle","该扩展专注于使用 LTX-2.3 进行 AI 视频生成，并提供图像编辑和画外音等支持任务的相关工具，保持了连贯的领域。",{"category":33,"check":37,"severity":24,"summary":38},"Description quality","显示的描述准确反映了扩展的功能，清楚地概述了其 AI 视频生成的目的和用例。",{"category":40,"check":41,"severity":24,"summary":42},"Invocation","Scoped tools","该扩展公开了 `ltx2.py` 等特定的、作用域限定的工具，其参数清晰定义，用于视频生成任务。",{"category":44,"check":45,"severity":24,"summary":46},"Documentation","Configuration & parameter reference","LTX-2.3 模型的所有参数都在 SKILL.md 中得到了清晰的文档记录，包括默认值和描述。",{"category":33,"check":48,"severity":24,"summary":49},"Tool naming","`ltx2.py` 和 `voiceover.py` 等工具名称具有描述性，并遵循通用约定。",{"category":33,"check":51,"severity":24,"summary":52},"Minimal I/O surface","`ltx2.py` 工具的输入参数特定于视频生成任务，输出是视频文件。",{"category":54,"check":55,"severity":24,"summary":56},"License","License usability","该扩展根据 MIT 许可证授权，这是一个宽松的开源许可证。",{"category":58,"check":59,"severity":24,"summary":60},"Maintenance","Commit recency","最后一次提交是在 2026 年 5 月 11 日，在过去 3 个月内。",{"category":58,"check":62,"severity":24,"summary":63},"Dependency Management","该项目包含用于 Python 依赖项的 `requirements.txt` 和用于 Node.js 依赖项的 `package.json`，表明对第三方库进行了管理。",{"category":65,"check":66,"severity":67,"summary":68},"Security","Secret Management","warning","SKILL.md 提到了 `.env` 中需要 `MODAL_LTX2_ENDPOINT_URL`，并且设置说明提到了 HuggingFace token 和 Gemma 3 许可证接受，但没有明确说明这些 secrets 如何被安全地处理或记录。",{"category":65,"check":70,"severity":24,"summary":71},"Injection","该技能主要使用结构化参数的直接工具调用，通过捆绑脚本减少了提示注入的风险。",{"category":65,"check":73,"severity":24,"summary":74},"Transitive Supply-Chain Grenades","该扩展捆绑了其依赖项和脚本，并且云 GPU 调用通过特定的工具接口进行管理，避免了运行时外部代码执行。",{"category":65,"check":76,"severity":24,"summary":77},"Sandbox Isolation","提供的脚本和工具在其定义的范围内运行，专注于视频生成任务，并且不尝试修改其预期目录之外的文件。",{"category":65,"check":79,"severity":24,"summary":80},"Sandbox escape primitives","在提供的脚本中未发现分离进程生成或拒绝重试循环的证据。",{"category":65,"check":82,"severity":24,"summary":83},"Data Exfiltration","该技能的主要功能是视频生成，没有迹象表明它会读取机密用户数据或将其提交给第三方。",{"category":65,"check":85,"severity":24,"summary":86},"Hidden Text Tricks","捆绑的内容似乎没有隐藏的操纵技巧，并且描述使用了标准的可见字符。",{"category":88,"check":89,"severity":24,"summary":90},"Hooks","Opaque code execution","捆绑的脚本是纯 Python 和 Node.js，没有混淆、base64 负载或运行时代码获取的证据。",{"category":92,"check":93,"severity":24,"summary":94},"Portability","Structural Assumption","设置说明和脚本示例假定标准的项目结构和命令行用法，提供了清晰的指导。",{"category":96,"check":97,"severity":24,"summary":98},"Trust","Issues Attention","在过去 90 天内，打开了 1 个问题，关闭了 4 个问题，表明维护者积极参与并且关闭率健康。",{"category":100,"check":101,"severity":24,"summary":102},"Versioning","Release Management","该项目有 GitHub releases，README 引用了版本信息，可以清晰地检测新版本。",{"category":104,"check":105,"severity":24,"summary":106},"Execution","Validation","Python 脚本会验证输入参数，如提示、宽度和高度，确保它们满足预期的格式和约束。",{"category":65,"check":108,"severity":24,"summary":109},"Unguarded Destructive Operations","主要操作涉及生成视频文件，这些文件不具有破坏性。云 GPU 的配置步骤是指导性的和明确的。",{"category":111,"check":112,"severity":24,"summary":113},"Code Execution","Error Handling","Python 脚本包含对参数和云操作的错误处理，文档概述了限制和恢复步骤。",{"category":111,"check":115,"severity":24,"summary":116},"Logging","设置说明和云部署脚本通过 Modal/RunPod 的基础设施暗示了日志记录，并且提到了 `--debug` 标志用于详细输出。",{"category":118,"check":119,"severity":120,"summary":121},"Compliance","GDPR","not_applicable","该扩展专注于视频生成，似乎不处理个人数据。",{"category":118,"check":123,"severity":24,"summary":124},"Target market","该扩展具有全球适用性，文档中未指明任何特定的区域或司法管辖区限制。",{"category":92,"check":126,"severity":24,"summary":127},"Runtime stability","该工具包指定了 Python 3.9+ 和 Node.js 18+ 的要求，并为云环境提供了清晰的设置说明，确保了广泛的兼容性。",{"category":44,"check":129,"severity":24,"summary":130},"README","README 非常全面，提供了安装说明、使用示例、功能和技术细节。",{"category":33,"check":132,"severity":24,"summary":133},"Tool surface size","该扩展公开了适量的 Python 工具和 Claude Code 命令，都与视频制作相关。",{"category":40,"check":135,"severity":24,"summary":136},"Overlapping near-synonym tools","工具名称和功能是不同的，涵盖了视频生成和制作的特定方面，没有明显的重叠。",{"category":44,"check":138,"severity":24,"summary":139},"Phantom features","所有宣传的功能，如 AI 视频生成、画外音和音乐，都有相应的工具和详细文档。",{"category":141,"check":142,"severity":24,"summary":143},"Install","Installation instruction","README 提供了清晰的分步安装说明，包括云 GPU 和 AI 模型的设置，以及可复制粘贴的示例。",{"category":145,"check":146,"severity":24,"summary":147},"Errors","Actionable error messages","文档概述了潜在的故障模式和恢复步骤，特别是对于云 GPU 设置和工具参数。",{"category":104,"check":149,"severity":24,"summary":150},"Pinned dependencies","依赖项通过 `requirements.txt` 和 `package.json` 进行管理，表明已固定版本以确保可重现性。",{"category":33,"check":152,"severity":120,"summary":153},"Dry-run preview","主要操作是视频生成，这在传统意义上不是一个有益于 dry-run 标志的状态更改操作。",{"category":155,"check":156,"severity":24,"summary":157},"Protocol","Idempotent retry & timeouts","通过 Modal 和 RunPod 的云操作通常会处理重试和超时，并且脚本结构表明每次生成任务都进行离散的、无状态的调用。",{"category":118,"check":159,"severity":24,"summary":160},"Telemetry opt-in","该扩展似乎默认不发送遥测数据；云提供商集成和可选的本地工具将管理自己的遥测策略。",{"category":40,"check":162,"severity":24,"summary":163},"Precise Purpose","该技能的目的精确定义为使用 LTX-2.3 进行 AI 视频生成，包括文本到视频和图像到视频，并具有清晰的用例和触发器。",{"category":40,"check":165,"severity":24,"summary":166},"Concise Frontmatter","SKILL.md 的 frontmatter 简洁明了，清楚地说明了核心功能并提供了相关的触发短语。",{"category":44,"check":168,"severity":24,"summary":169},"Concise Body","SKILL.md 结构良好，避免了过多的冗长，将更深入的内容委托给单独的文件，如 `references` 和 `examples`。",{"category":171,"check":172,"severity":24,"summary":173},"Context","Progressive Disclosure","文档有效地使用了渐进式披露，主 SKILL.md 概述了过程，并将详细信息链接到单独的文件。",{"category":171,"check":175,"severity":120,"summary":176},"Forked exploration","该技能专注于生成，不涉及需要 `context: fork` 的深度代码审查或探索。",{"category":22,"check":178,"severity":24,"summary":179},"Usage examples","SKILL.md 和 README 提供了许多清晰、可直接使用的示例，用于文本到视频、图像到视频和各种参数，展示了可观察到的结果。",{"category":22,"check":181,"severity":24,"summary":182},"Edge cases","SKILL.md 文档提到了已知的限制，如文本渲染问题和最大持续时间，并提出了恢复建议（例如，不同的种子、拼接）。",{"category":111,"check":184,"severity":24,"summary":185},"Tool Fallback","该扩展允许可选的云 GPU 使用 Modal 或 RunPod，并且如果未配置云端点，`ltx2.py` 工具也可以在本地运行，从而提供了回退选项。",{"category":92,"check":187,"severity":24,"summary":188},"Stack assumptions","该项目清楚地声明了 Node.js 和 Python 版本的需求，并为 GPU 任务指定了云环境（Modal/RunPod）。",{"category":190,"check":191,"severity":24,"summary":192},"Safety","Halt on unexpected state","设置说明引导用户完成配置步骤，暗示如果未满足先决条件，工作流程将以清晰的错误停止。",{"category":92,"check":194,"severity":24,"summary":195},"Cross-skill coupling","该技能是独立的，专注于视频生成，没有隐含地依赖于同时加载的其他特定技能。",1778686374491,"该技能支持使用 LTX-2.3 22B DiT 模型进行 AI 视频生成，支持文本到视频和图像到视频片段的创建。它提供了详细的参数控制、样式 LoRA，并通过 Modal 或 RunPod 与云 GPU 集成。",[199,200,201,202,203,204],"文本到视频生成","图像到视频动画","可自定义的视频参数（分辨率、时长、质量）","用于美学控制的样式 LoRA 应用","与云 GPU 平台集成（Modal、RunPod）","详细的提示指南和用例示例",[206,207,208,209],"生成精度至关重要的逼真人物唇语同步视频","生成超过约 8 秒的视频而不进行拼接","直接在视频帧中生成可读文本","直接提供画外音或音乐（需要与其他工具集成）","3.0.0","4.4.0","从文本提示或现有图像生成 AI 驱动的视频片段，以满足各种视频制作需求，例如创建备用镜头、动画背景或动态内容。","该扩展得分很高，因为它拥有全面的文档、清晰的示例和明确定义的范围。关于 `.env` 文件中 secret 管理处理的轻微警告使其无法获得满分。",96,"优秀的 AI 视频生成技能，文档和示例非常详尽。",[217,218,219,220,221,222,223],"video-generation","ltx-2-3","text-to-video","image-to-video","modal","runpod","ai-video","community",[226,227,228,229,230],"生成氛围备用镜头片段","为动态内容制作静态图像动画","创建动画幻灯片背景","生成风格化角色客串","制作品牌片头/片尾动态背景",{"codeQuality":232,"collectedAt":234,"documentation":235,"maintenance":238,"security":244,"testCoverage":247},{"hasLockfile":233},true,1778686338096,{"descriptionLength":236,"readmeSize":237},314,18886,{"closedIssues90d":239,"forks":240,"hasChangelog":233,"openIssues90d":241,"pushedAt":242,"stars":243},4,187,1,1778501732000,1137,{"hasNpmPackage":245,"license":246,"smitheryVerified":245},false,"MIT",{"hasCi":233,"hasTests":245},{"updatedAt":249},1778686575608,{"basePath":251,"githubOwner":252,"githubRepo":253,"locale":18,"slug":13,"type":254},".claude/skills/ltx2","digitalsamba","claude-code-video-toolkit","skill",null,{"evaluate":257,"extract":260},{"promptVersionExtension":210,"promptVersionScoring":211,"score":214,"tags":258,"targetMarket":259,"tier":224},[217,218,219,220,221,222,223],"global",{"commitSha":261},"HEAD",{"repoId":263,"translatedFrom":264},"kd70r97eght58pp9f1x8scdagd86n32q","k174cmwts8pcx7rj20ryz4xf4n86nqrg",{"_creationTime":266,"_id":263,"identity":267,"providers":268,"workflow":397},1778686211924.9185,{"githubOwner":252,"githubRepo":253,"sourceUrl":14},{"classify":269,"discover":380,"github":383},{"commitSha":261,"extensions":270},[271,283,294,303,311,317,325,334,348,357,365,372],{"basePath":272,"description":273,"displayName":274,"installMethods":275,"rationale":276,"selectedPaths":277,"source":281,"sourceLanguage":282,"type":254},".claude/skills/acestep","AI music generation with ACE-Step 1.5 — background music, vocal tracks, covers, stem extraction, audio repainting, and continuation for video production. Use when generating music, soundtracks, jingles, or working with audio stems. Triggers include background music, soundtrack, jingle, music generation, stem extraction, cover, style transfer, repaint, continuation, or musical composition tasks.","acestep",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/acestep/SKILL.md",[278],{"path":279,"priority":280},"SKILL.md","mandatory","rule","en",{"basePath":284,"description":285,"displayName":286,"installMethods":287,"rationale":288,"selectedPaths":289,"source":281,"sourceLanguage":282,"type":254},".claude/skills/elevenlabs","Generate AI voiceovers, sound effects, and music using ElevenLabs APIs. Use when creating audio content for videos, podcasts, or games. Triggers include generating voiceovers, narration, dialogue, sound effects from descriptions, background music, soundtrack generation, voice cloning, or any audio synthesis task.","elevenlabs",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/elevenlabs/SKILL.md",[290,291],{"path":279,"priority":280},{"path":292,"priority":293},"reference.md","medium",{"basePath":295,"description":296,"displayName":297,"installMethods":298,"rationale":299,"selectedPaths":300,"source":281,"sourceLanguage":282,"type":254},".claude/skills/ffmpeg","Video and audio processing with FFmpeg. Use for format conversion, resizing, compression, audio extraction, and preparing assets for Remotion. Triggers include converting GIF to MP4, resizing video, extracting audio, compressing files, or any media transformation task.","ffmpeg",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/ffmpeg/SKILL.md",[301,302],{"path":279,"priority":280},{"path":292,"priority":293},{"basePath":304,"description":305,"displayName":306,"installMethods":307,"rationale":308,"selectedPaths":309,"source":281,"sourceLanguage":282,"type":254},".claude/skills/frontend-design","Create distinctive, production-grade frontend interfaces with high design quality. Use this skill when the user asks to build web components, pages, or applications. Generates creative, polished code that avoids generic AI aesthetics.","frontend-design",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/frontend-design/SKILL.md",[310],{"path":279,"priority":280},{"basePath":251,"description":312,"displayName":13,"installMethods":313,"rationale":314,"selectedPaths":315,"source":281,"sourceLanguage":282,"type":254},"AI video generation with LTX-2.3 22B — text-to-video, image-to-video clips for video production. Use when generating video clips, animating images, creating b-roll, animated backgrounds, or motion content. Triggers include video generation, animate image, b-roll, motion, video clip, text-to-video, image-to-video.",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/ltx2/SKILL.md",[316],{"path":279,"priority":280},{"basePath":318,"description":319,"displayName":320,"installMethods":321,"rationale":322,"selectedPaths":323,"source":281,"sourceLanguage":282,"type":254},".claude/skills/moviepy","Python video composition with moviepy 2.x — overlaying deterministic text on AI-generated video (LTX-2, SadTalker), compositing clips, single-file build.py video projects. Use when adding labels/captions/lower-thirds to LTX-2 or SadTalker outputs, building short ad-style spots in pure Python without Remotion, or doing programmatic video composition. Triggers include text overlay on video, label LTX-2 clip, caption SadTalker output, lower third, build.py video, moviepy, Python video composition, sub-30s ad spot.","moviepy",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/moviepy/SKILL.md",[324],{"path":279,"priority":280},{"basePath":326,"description":327,"displayName":328,"installMethods":329,"rationale":330,"selectedPaths":331,"source":281,"sourceLanguage":282,"type":254},".claude/skills/playwright-recording","Record browser interactions as video using Playwright. Use for capturing demo videos, app walkthroughs, and UI flows for Remotion videos. Triggers include recording a demo, capturing browser video, screen recording a website, or creating walkthrough footage.","playwright-recording",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/playwright-recording/SKILL.md",[332,333],{"path":279,"priority":280},{"path":292,"priority":293},{"basePath":335,"description":336,"displayName":337,"installMethods":338,"rationale":339,"selectedPaths":340,"source":281,"sourceLanguage":282,"type":254},".claude/skills/qwen-edit","AI image editing prompting patterns for Qwen-Image-Edit. Use when editing photos while preserving identity, reframing cropped images, changing clothing or accessories, adjusting poses, applying style transfers, or character transformations. Provides prompt patterns, parameter tuning, and examples.","qwen-edit",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/qwen-edit/SKILL.md",[341,342,344,346],{"path":279,"priority":280},{"path":343,"priority":293},"examples.md",{"path":345,"priority":293},"parameters.md",{"path":347,"priority":293},"prompting.md",{"basePath":349,"description":350,"displayName":351,"installMethods":352,"rationale":353,"selectedPaths":354,"source":281,"sourceLanguage":282,"type":254},".claude/skills/remotion","Toolkit-specific Remotion patterns — custom transitions, shared components, and project conventions. For core Remotion framework knowledge (hooks, animations, rendering, etc.), see the `remotion-official` skill.","remotion",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/remotion/SKILL.md",[355,356],{"path":279,"priority":280},{"path":292,"priority":293},{"basePath":358,"description":359,"displayName":360,"installMethods":361,"rationale":362,"selectedPaths":363,"source":281,"sourceLanguage":282,"type":254},".claude/skills/remotion-official","Best practices for Remotion - Video creation in React","remotion-best-practices",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/remotion-official/SKILL.md",[364],{"path":279,"priority":280},{"basePath":366,"description":367,"displayName":222,"installMethods":368,"rationale":369,"selectedPaths":370,"source":281,"sourceLanguage":282,"type":254},".claude/skills/runpod","Cloud GPU processing via RunPod serverless. Use when setting up RunPod endpoints, deploying Docker images, managing GPU resources, troubleshooting endpoint issues, or understanding costs. Covers all 5 toolkit images (qwen-edit, realesrgan, propainter, sadtalker, qwen3-tts).",{"claudeCode":12},"SKILL.md frontmatter at .claude/skills/runpod/SKILL.md",[371],{"path":279,"priority":280},{"basePath":373,"description":374,"displayName":375,"installMethods":376,"rationale":377,"selectedPaths":378,"source":281,"sourceLanguage":282,"type":254},"skills/openclaw-video-toolkit","Create professional videos autonomously using claude-code-video-toolkit — AI voiceovers, image generation, music, talking heads, and Remotion rendering.","openclaw-video-toolkit",{"claudeCode":12},"SKILL.md frontmatter at skills/openclaw-video-toolkit/SKILL.md",[379],{"path":279,"priority":280},{"sources":381},[382],"manual",{"closedIssues90d":239,"description":384,"forks":240,"license":246,"openIssues90d":241,"pushedAt":242,"readmeSize":237,"stars":243,"topics":385},"AI-native video production toolkit for Claude Code",[386,387,388,286,389,390,351,391,392,393,394,395,396],"ai-video-generator","claude-code","developer-tools","playwright","programmatic-video","text-to-speech","video-editing","video-production","open-source","qwen-tts","openclaw",{"classifiedAt":398,"discoverAt":399,"extractAt":400,"githubAt":400,"updatedAt":398},1778686219532,1778686211925,1778686217771,[223,220,218,221,222,219,217],{"evaluatedAt":403,"extractAt":404,"updatedAt":249},1778686374605,1778686219732,[],[407,437,460,491,519,550],{"_creationTime":408,"_id":409,"community":410,"display":411,"identity":417,"providers":421,"relations":429,"tags":432,"workflow":433},1778695495612.4958,"k173ff9fw8mhej6fd22bqt537x86mvc6",{"reviewCount":8},{"description":412,"installMethods":413,"name":415,"sourceUrl":416},"从文本或图像生成短的 AI 视频——文本到视频、图像到视频和基于参考的生成——无需设置 API 密钥。当用户想要创建视频片段、制作图像动画或根据描述生成视频时使用。\n",{"claudeCode":414},"pexoai/pexo-skills","videoagent-video-studio","https://github.com/pexoai/pexo-skills",{"basePath":418,"githubOwner":419,"githubRepo":420,"locale":18,"slug":415,"type":254},"skills/videoagent-video-studio","pexoai","pexo-skills",{"evaluate":422,"extract":428},{"promptVersionExtension":210,"promptVersionScoring":211,"score":423,"tags":424,"targetMarket":259,"tier":427},100,[425,223,219,220,426],"video","generation","verified",{"commitSha":261},{"repoId":430,"translatedFrom":431},"kd7a11fjbvhz4653zn8wdan1bd86mrja","k17ebhanfep7e674dkqscx6dt586mf11",[223,426,220,219,425],{"evaluatedAt":434,"extractAt":435,"updatedAt":436},1778695435906,1778695321983,1778695495612,{"_creationTime":438,"_id":439,"community":440,"display":441,"identity":444,"providers":445,"relations":454,"tags":456,"workflow":457},1778686643435.0105,"k17ezar195r33hy4qrk2q42kdn86m3r4",{"reviewCount":8},{"description":442,"installMethods":443,"name":375,"sourceUrl":14},"使用 claude-code-video-toolkit 自主创建专业视频 — AI 配音、图像生成、音乐、谈话头像和 Remotion 渲染。",{"claudeCode":12},{"basePath":373,"githubOwner":252,"githubRepo":253,"locale":18,"slug":375,"type":254},{"evaluate":446,"extract":453},{"promptVersionExtension":210,"promptVersionScoring":211,"score":447,"tags":448,"targetMarket":259,"tier":427},99,[217,449,351,221,450,451,452],"ai-tools","python","typescript","automation",{"commitSha":261},{"repoId":263,"translatedFrom":455},"k17crq1jtpdq6v8et37enw77pd86mtmh",[449,452,221,450,351,451,217],{"evaluatedAt":458,"extractAt":404,"updatedAt":459},1778686517002,1778686643435,{"_creationTime":461,"_id":462,"community":463,"display":464,"identity":470,"providers":474,"relations":483,"tags":486,"workflow":487},1778697578571.0342,"k1746jwd0qyxc70wwdspb0705x86m5dy",{"reviewCount":8},{"description":465,"installMethods":466,"name":468,"sourceUrl":469},"起草和优化视频生成模型（文本到视频和图像到视频）的提示，并在目标是角色一致性后再进行图像到视频生成时，为图像模型创建角色表提示。当用户请求“视频提示”、特定模型提示（如 Seedance 2.0、Ovi、Sora、Veo 3、Wan 2.2、LTX-2 或 LTX-2.3）或一致性角色提示（如“角色表提示”、“角色轮廓”、“角色参考表”或“照片身份表”）时使用。",{"claudeCode":467},"Square-Zero-Labs/video-prompting-skill","Video Prompting Skill","https://github.com/Square-Zero-Labs/video-prompting-skill",{"basePath":471,"githubOwner":472,"githubRepo":473,"locale":18,"slug":471,"type":254},"video-prompting","Square-Zero-Labs","video-prompting-skill",{"evaluate":475,"extract":481},{"promptVersionExtension":210,"promptVersionScoring":211,"score":476,"tags":477,"targetMarket":259,"tier":427},95,[217,478,479,480,219,220],"prompt-engineering","character-consistency","ai-art",{"commitSha":261,"license":482},"Apache-2.0",{"repoId":484,"translatedFrom":485},"kd7a3nw9bpm9b0jdnkskwdnrvx86ndrj","k1778wrb4sjw606cz318m6fc9d86mg0e",[480,479,220,478,219,217],{"evaluatedAt":488,"extractAt":489,"updatedAt":490},1778697559479,1778697542749,1778697578571,{"_creationTime":492,"_id":493,"community":494,"display":495,"identity":500,"providers":504,"relations":512,"tags":515,"workflow":516},1778685615701.8425,"k1707ctze9p8fn1e339nd2czjn86m6p5",{"reviewCount":8},{"description":496,"installMethods":497,"name":425,"sourceUrl":499},"When the user wants to create, generate, or produce video content using AI tools or programmatic frameworks. Also use when the user mentions 'video production,' 'AI video,' 'Remotion,' 'Hyperframes,' 'HeyGen,' 'Synthesia,' 'Veo,' 'Runway,' 'Kling,' 'Pika,' 'video generation,' 'AI avatar,' 'talking head video,' 'programmatic video,' 'video template,' 'explainer video,' 'product demo video,' 'video pipeline,' or 'make me a video.' Use this for video creation, generation, and production workflows. For video content strategy and what to post, see social-content. For paid video ad creative, see ad-creative.",{"claudeCode":498},"coreyhaines31/marketingskills","https://github.com/coreyhaines31/marketingskills",{"basePath":501,"githubOwner":502,"githubRepo":503,"locale":282,"slug":425,"type":254},"skills/video","coreyhaines31","marketingskills",{"evaluate":505,"extract":511},{"promptVersionExtension":210,"promptVersionScoring":211,"score":423,"tags":506,"targetMarket":259,"tier":427},[425,507,223,351,508,509,510],"production","heygen","runway","content-creation",{"commitSha":261},{"parentExtensionId":513,"repoId":514},"k175jvka8cxxkf91gk8qy25r8186npjr","kd7a4vjty5ay3s25r82cm72wdn86nmg0",[223,510,508,507,351,509,425],{"evaluatedAt":517,"extractAt":518,"updatedAt":517},1778686582142,1778685615701,{"_creationTime":520,"_id":521,"community":522,"display":523,"identity":529,"providers":533,"relations":543,"tags":546,"workflow":547},1778695548458.3403,"k17874jrpzpfq3ftmmahe017n586nmy2",{"reviewCount":8},{"description":524,"installMethods":525,"name":527,"sourceUrl":528},"Compose or analyze sacred music in Hildegard von Bingen's distinctive modal style. Covers modal selection, melodic contour (wide-range melodies), text-setting (syllabic and melismatic), neumatic notation, and liturgical context for antiphons, sequences, and responsories. Use when composing a new piece in Hildegardian style, analyzing an existing chant for structure and mode, researching medieval modal music, preparing to perform or teach Hildegard's music, or setting Latin sacred texts.\n",{"claudeCode":526},"pjt222/agent-almanac","compose-sacred-music","https://github.com/pjt222/agent-almanac",{"basePath":530,"githubOwner":531,"githubRepo":532,"locale":282,"slug":527,"type":254},"skills/compose-sacred-music","pjt222","agent-almanac",{"evaluate":534,"extract":542},{"promptVersionExtension":210,"promptVersionScoring":211,"score":423,"tags":535,"targetMarket":259,"tier":427},[536,537,538,221,539,540,541],"hildegard","sacred-music","chant","musicology","composition","analysis",{"commitSha":261},{"parentExtensionId":544,"repoId":545},"k170h0janaa9kwn7cfgfz2ykss86mmh9","kd7aryv63z61j39n2td1aeqkvh86mh12",[541,538,540,536,221,539,537],{"evaluatedAt":548,"extractAt":549,"updatedAt":548},1778696688896,1778695548458,{"_creationTime":551,"_id":552,"community":553,"display":554,"identity":559,"providers":562,"relations":569,"tags":572,"workflow":573},1778691530223.3884,"k1718hxj2cjyjrmhxfbj7theex86mdba",{"reviewCount":8},{"description":555,"installMethods":556,"name":396,"sourceUrl":558},"使用多提供商路由从文本生成图像和视频 — 支持 GPT Image 2.0（近乎完美的文本渲染）、Nanobanana 2、Seedream 5.0、Midjourney V8.1（统一的逼真+动漫风格）、Flux 2 Klein（经济高效的草稿）、Seedance 2.0 / Happyhorse 1.0 / Veo 3.1 视频，以及本地 ComfyUI 工作流。包含 1,446 个精选提示和风格感知提示增强。当用户想要创建图像/视频、设计素材、制作照片动画、增强提示或管理 AI 艺术工作流时使用。不适用于：通用聊天、代码生成、文档编写、现有素材的视频编辑、音频/TTS，或任何与 AI 图像/视频创建无关的任务。",{"claudeCode":557},"jau123/MeiGen-AI-Design-MCP","https://github.com/jau123/MeiGen-AI-Design-MCP",{"basePath":396,"githubOwner":560,"githubRepo":561,"locale":18,"slug":396,"type":254},"jau123","MeiGen-AI-Design-MCP",{"evaluate":563,"extract":568},{"promptVersionExtension":210,"promptVersionScoring":211,"score":423,"tags":564,"targetMarket":259,"tier":427},[565,217,480,566,478,567],"image-generation","comfyui","design-assistant",{"commitSha":261},{"repoId":570,"translatedFrom":571},"kd7b1trrphkh2efj71w89w7qzh86nh22","k175txezbab4wcy00m9gcdq8qx86mz9k",[480,566,567,565,478,217],{"evaluatedAt":574,"extractAt":575,"updatedAt":576},1778691399174,1778691316210,1778691530223]