[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"extension-skill-answerzhao-tts-vi":3,"guides-for-answerzhao-tts":222,"similar-k1712xyy3wyvy83c0f9z7kccg9866jg4":223},{"_creationTime":4,"_id":5,"children":6,"community":7,"display":9,"evaluation":20,"identity":189,"isFallback":193,"parentExtension":194,"providers":195,"relations":199,"repo":201,"workflow":219},1778054691785.2524,"k1712xyy3wyvy83c0f9z7kccg9866jg4",[],{"reviewCount":8},0,{"description":10,"installMethods":11,"name":12,"sourceUrl":13,"tags":14},"Implement text-to-speech (TTS) capabilities using the z-ai-web-dev-sdk. Use this skill when the user needs to convert text into natural-sounding speech, create audio content, build voice-enabled applications, or generate spoken audio files. Supports multiple voices, adjustable speed, and various audio formats.",{},"Text-to-Speech (TTS)","https://github.com/answerzhao/agent-skills/tree/HEAD/glm-skills/TTS",[15,16,17,18,19],"tts","text-to-speech","audio","sdk","z-ai-web-dev-sdk",{"_creationTime":21,"_id":22,"extensionId":5,"locale":23,"result":24,"trustSignals":177,"workflow":187},1778054738050.0125,"kn72d4s976zsvx6r9sy8r1qax98664ee","en",{"checks":25,"evaluatedAt":167,"extensionSummary":168,"promptVersionExtension":169,"promptVersionScoring":170,"rationale":171,"score":172,"summary":173,"tags":174,"targetMarket":175,"tier":176},[26,31,34,37,41,44,48,52,55,58,62,67,70,74,77,80,83,86,89,92,96,100,104,109,113,116,119,122,126,129,132,135,138,141,145,148,151,154,157,160,164],{"category":27,"check":28,"severity":29,"summary":30},"Practical Utility","Problem relevance","pass","The description clearly names the problem of converting text into natural-sounding speech and lists specific use cases like creating audio content and building voice-enabled applications.",{"category":27,"check":32,"severity":29,"summary":33},"Unique selling proposition","The skill offers unique TTS capabilities beyond a simple API wrapper by integrating with the z-ai-web-dev-sdk, providing control over voices, speed, and formats, which offers value over basic prompting.",{"category":27,"check":35,"severity":29,"summary":36},"Production readiness","The skill provides comprehensive documentation including CLI usage, SDK implementation examples, and best practices, covering the full lifecycle for TTS generation.",{"category":38,"check":39,"severity":29,"summary":40},"Scope","Single responsibility principle","The skill focuses solely on Text-to-Speech functionality, clearly named and described, without extending into unrelated domains.",{"category":38,"check":42,"severity":29,"summary":43},"Description quality","The description is accurate, concise, readable, and accurately reflects the skill's TTS capabilities and its use of the z-ai-web-dev-sdk.",{"category":45,"check":46,"severity":29,"summary":47},"Invocation","Scoped tools","The skill uses a single, well-defined tool (`zai.audio.tts.create`) for its core functionality, avoiding a generalist command.",{"category":49,"check":50,"severity":29,"summary":51},"Documentation","Configuration & parameter reference","All parameters for the TTS functionality, including input text constraints, audio parameters (speed, volume), and format options, are clearly documented with their ranges and defaults.",{"category":38,"check":53,"severity":29,"summary":54},"Tool naming","The primary tool used, `zai.audio.tts.create`, is descriptive within the context of the SDK.",{"category":38,"check":56,"severity":29,"summary":57},"Minimal I/O surface","The TTS tool's input parameters (text, voice, speed, format, stream) are specific to the task, and the output is the audio data, fulfilling the promised payload without extraneous information.",{"category":59,"check":60,"severity":29,"summary":61},"License","License usability","The extension includes a LICENSE.txt file with the MIT license, which is a permissive open-source license.",{"category":63,"check":64,"severity":65,"summary":66},"Maintenance","Commit recency","not_applicable","No commits are available for evaluation as the repository is not being tracked for commits in this context.",{"category":63,"check":68,"severity":65,"summary":69},"Dependency Management","The skill relies on the `z-ai-web-dev-sdk`, which is assumed to be managed externally; no other third-party dependencies are bundled or managed within the skill's repository.",{"category":71,"check":72,"severity":29,"summary":73},"Security","Secret Management","The skill itself does not appear to handle or expose secrets directly. The `z-ai-web-dev-sdk` is assumed to handle its own authentication securely.",{"category":71,"check":75,"severity":29,"summary":76},"Injection","The skill's code and documentation do not show any explicit or implicit handling of untrusted external data as instructions.",{"category":71,"check":78,"severity":29,"summary":79},"Transitive Supply-Chain Grenades","The skill does not fetch external files at runtime or execute remote scripts, keeping all content within the bundled repository.",{"category":71,"check":81,"severity":29,"summary":82},"Sandbox Isolation","The skill's operations are confined to generating audio files and do not involve modifying files outside its designated scope or the project folder.",{"category":71,"check":84,"severity":29,"summary":85},"Sandbox escape primitives","No detached-process spawns or deny-retry loops were found in the skill's scripts.",{"category":71,"check":87,"severity":29,"summary":88},"Data Exfiltration","The skill does not make any undocumented outbound calls or attempt to exfiltrate confidential data. All interactions are with the specified z-ai-web-dev-sdk.",{"category":71,"check":90,"severity":29,"summary":91},"Hidden Text Tricks","The bundled files and descriptions are free of hidden-steering tricks, invisible characters, or other obfuscation methods.",{"category":93,"check":94,"severity":29,"summary":95},"Hooks","Opaque code execution","The skill's code is written in plain, readable JavaScript and does not employ obfuscation techniques like base64 payloads or minified bundles without source maps.",{"category":97,"check":98,"severity":29,"summary":99},"Portability","Structural Assumption","The skill does not make structural assumptions about the user's project organization, using standard file operations and SDK interactions.",{"category":101,"check":102,"severity":65,"summary":103},"Trust","Issues Attention","No issue data is available for evaluation.",{"category":105,"check":106,"severity":107,"summary":108},"Versioning","Release Management","warning","No manifest version, GitHub release tags, or CHANGELOG are present, and install instructions are not provided, making it difficult to track versions.",{"category":110,"check":111,"severity":29,"summary":112},"Code Execution","Validation","Input parameters like text length, speed, and volume ranges are validated according to API constraints, and the output is the audio data, implying proper handling.",{"category":71,"check":114,"severity":29,"summary":115},"Unguarded Destructive Operations","The skill is read-only in nature, focusing on audio generation, and does not perform any destructive operations.",{"category":110,"check":117,"severity":29,"summary":118},"Error Handling","The provided JavaScript examples include try-catch blocks for handling potential errors during TTS generation and file writing, reporting meaningful messages.",{"category":110,"check":120,"severity":29,"summary":121},"Logging","The example `tts.ts` script logs success or failure messages to the console, providing a basic audit trail of its execution.",{"category":123,"check":124,"severity":65,"summary":125},"Compliance","GDPR","The skill does not process personal data; it converts text to speech.",{"category":123,"check":127,"severity":29,"summary":128},"Target market","The skill operates on text and audio generation, with no regional or jurisdictional logic, and therefore targets a global market.",{"category":97,"check":130,"severity":29,"summary":131},"Runtime stability","The skill is written in standard JavaScript and relies on the Node.js environment and the `z-ai-web-dev-sdk`, with no specific OS or shell assumptions made.",{"category":45,"check":133,"severity":29,"summary":134},"Precise Purpose","The description clearly states the purpose of implementing text-to-speech using the z-ai-web-dev-sdk and lists specific use cases, making its function and target audience clear.",{"category":45,"check":136,"severity":29,"summary":137},"Concise Frontmatter","The frontmatter is concise and self-contained, accurately summarizing the core capability and its purpose.",{"category":49,"check":139,"severity":29,"summary":140},"Concise Body","The SKILL.md body is well-structured and uses progressive disclosure, keeping the main instruction set concise.",{"category":142,"check":143,"severity":29,"summary":144},"Context","Progressive Disclosure","The SKILL.md outlines the flow and provides examples, with detailed information and API constraints clearly separated. The split text function is provided inline, which is reasonable for its size.",{"category":142,"check":146,"severity":65,"summary":147},"Forked exploration","This skill is not an exploration-focused skill; it performs a direct action (TTS generation) and does not require forked exploration.",{"category":27,"check":149,"severity":29,"summary":150},"Usage examples","The skill provides numerous end-to-end examples for CLI usage, basic SDK implementation, batch processing, and integration examples, covering various scenarios.",{"category":27,"check":152,"severity":29,"summary":153},"Edge cases","The documentation addresses edge cases such as text length limits, speed/volume parameter ranges, streaming format limitations, and provides solutions and error handling advice.",{"category":110,"check":155,"severity":65,"summary":156},"Tool Fallback","The skill directly uses the `z-ai-web-dev-sdk` and does not rely on external tools like MCP that would require fallbacks or version declarations.",{"category":97,"check":158,"severity":29,"summary":159},"Stack assumptions","The skill specifies its JavaScript/Node.js environment and the dependency on `z-ai-web-dev-sdk`, and provides clear examples for integration.",{"category":161,"check":162,"severity":29,"summary":163},"Safety","Halt on unexpected state","The provided examples include robust error handling with try-catch blocks and explicit checks for input validity, ensuring the workflow halts on unexpected states.",{"category":97,"check":165,"severity":29,"summary":166},"Cross-skill coupling","The skill is self-contained and focuses solely on TTS functionality, without any implicit reliance on other skills.",1778054707160,"This skill enables text-to-speech conversion via the z-ai-web-dev-sdk. It provides examples for CLI usage, SDK integration, batch processing, and dynamic content generation, with controls for voice, speed, and format. It also includes best practices for text preparation, error handling, and SDK instance reuse.","2.0.0","3.4.0","The extension is well-documented with clear usage examples for both CLI and SDK. It adheres to best practices for error handling and security, and its scope is well-defined. The only minor issue is the lack of versioning information, which is a warning but does not detract significantly from the overall quality.",95,"A comprehensive text-to-speech skill utilizing the z-ai-web-dev-sdk, offering extensive documentation and examples.",[15,16,17,18,19],"global","verified",{"codeQuality":178,"collectedAt":179,"documentation":180,"maintenance":182,"security":183,"testCoverage":186},{},1778054695723,{"descriptionLength":181,"readmeSize":8},311,{},{"hasNpmPackage":184,"license":185,"smitheryVerified":184},false,"MIT",{"hasCi":184,"hasTests":184},{"updatedAt":188},1778054738050,{"githubOwner":190,"githubRepo":191,"locale":23,"slug":15,"type":192},"answerzhao","agent-skills","skill",true,null,{"extract":196,"llm":198},{"commitSha":197,"license":185},"aad73edbd0d9ffbc3d6a402b6eafa6dab96d5ebb",{"promptVersionExtension":169,"promptVersionScoring":170,"score":172,"targetMarket":175,"tier":176},{"repoId":200},"kd712v2g1pay70swwj0jpv2ggs864zgh",{"_creationTime":202,"_id":200,"identity":203,"providers":205,"workflow":216},1777995558409.901,{"githubOwner":190,"githubRepo":191,"sourceUrl":204},"https://github.com/answerzhao/agent-skills",{"discover":206,"github":209},{"sources":207},[208],"skills-sh",{"closedIssues90d":8,"forks":210,"openIssues90d":211,"pushedAt":212,"readmeSize":213,"stars":214,"topics":215},15,1,1768478800000,770,26,[],{"discoverAt":217,"extractAt":218,"githubAt":218,"updatedAt":218},1777995558409,1778054693420,{"anyEnrichmentAt":220,"extractAt":221,"githubAt":220,"llmAt":188,"updatedAt":188},1778054692243,1778054691785,[],[224,259,284,304,323,351],{"_creationTime":225,"_id":226,"community":227,"display":228,"identity":243,"providers":247,"relations":252,"workflow":255},1778053197391.382,"k170wvt5rx3c1hv9a5sfkyezc1866k0q",{"reviewCount":8},{"description":229,"installMethods":230,"name":231,"sourceUrl":232,"tags":233},"Universal AI voice / text-to-speech skill supporting OpenAI TTS (gpt-4o-mini-tts, tts-1), ElevenLabs multilingual TTS with voice cloning, Bailian Qwen TTS (qwen-tts / qwen3-tts-vd with voice-design custom voices, long-text chunking built in), MiniMax speech-02-hd, SiliconFlow CosyVoice / SenseVoice, and PlayHT 2.0. Use this skill whenever the user asks to read text aloud, synthesize speech, generate narration, create voice-over, dub a script, or turn any text into audio (mp3 / wav / ogg / flac). Typical phrases include \"read this aloud\", \"generate voice for ...\", \"create a narration of ...\", \"tts this\", \"把这段念出来\", \"做个配音\", \"合成语音\", or mentions of voices / TTS model names like Alloy, Ash, Cherry, Rachel, CosyVoice, PlayHT. Always use this skill even if the user does not specify a provider — pick one from EXTEND.md defaults or available env keys.",{},"Happy Audio Gen","https://github.com/iamzhihuix/happy-claude-skills/tree/HEAD/skills/happy-audio-gen",[15,234,17,235,236,237,238,239,240,241,242],"speech","voice-generation","openai","elevenlabs","bailian","minimax","siliconflow","playht","bun",{"githubOwner":244,"githubRepo":245,"locale":23,"slug":246,"type":192},"iamzhihuix","happy-claude-skills","happy-audio-gen",{"extract":248,"llm":250},{"commitSha":249,"license":185},"f49e7782a551759c9f9e0a4d4417ff053f0a86fd",{"promptVersionExtension":169,"promptVersionScoring":170,"score":251,"targetMarket":175,"tier":176},100,{"parentExtensionId":253,"repoId":254},"k173ydbbp6c0vdpxv5r0q9yvgd867en5","kd7dbbtdq95nkcs3k7fg9w6fdn864j0b",{"anyEnrichmentAt":256,"extractAt":257,"githubAt":256,"llmAt":258,"updatedAt":258},1778053199195,1778053197391,1778053284450,{"_creationTime":260,"_id":261,"community":262,"display":263,"identity":271,"providers":273,"relations":278,"workflow":280},1778053440456.6604,"k17a2cxtswmmk54b8wmpfbp5f9866jr0",{"reviewCount":8},{"description":264,"installMethods":265,"name":266,"sourceUrl":267,"tags":268},"Convert text to speech using ElevenLabs voice AI. Use when generating audio from text, creating voiceovers, building voice apps, or synthesizing speech in 70+ languages.",{},"ElevenLabs Text-to-Speech","https://github.com/elevenlabs/skills/tree/HEAD/text-to-speech",[16,237,269,17,270],"voice","synthesis",{"githubOwner":237,"githubRepo":272,"locale":23,"slug":16,"type":192},"skills",{"extract":274,"llm":276},{"commitSha":275,"license":185},"b476f0ccf4be0e22b2e77cc39307665425d1472b",{"promptVersionExtension":169,"promptVersionScoring":170,"score":277,"targetMarket":175,"tier":176},98,{"repoId":279},"kd71z3hz1pg97d1k2d6kaqeqtx864knt",{"anyEnrichmentAt":281,"extractAt":282,"githubAt":281,"llmAt":283,"updatedAt":283},1778053440833,1778053440456,1778053480675,{"_creationTime":285,"_id":286,"community":287,"display":288,"identity":298,"providers":299,"relations":302,"workflow":303},1778054691785.2515,"k17ev68gbw25zazp0w5z2a61hd8662cc",{"reviewCount":8},{"description":289,"installMethods":290,"name":291,"sourceUrl":292,"tags":293},"Implement speech-to-text (ASR/automatic speech recognition) capabilities using the z-ai-web-dev-sdk. Use this skill when the user needs to transcribe audio files, convert speech to text, build voice input features, or process audio recordings. Supports base64 encoded audio files and returns accurate text transcriptions.",{},"ASR (Speech to Text) Skill","https://github.com/answerzhao/agent-skills/tree/HEAD/glm-skills/ASR",[294,295,296,18,297,17],"asr","speech-to-text","transcription","cli",{"githubOwner":190,"githubRepo":191,"locale":23,"slug":294,"type":192},{"extract":300,"llm":301},{"commitSha":197,"license":185},{"promptVersionExtension":169,"promptVersionScoring":170,"score":172,"targetMarket":175,"tier":176},{"repoId":200},{"anyEnrichmentAt":220,"extractAt":221,"githubAt":220,"llmAt":188,"updatedAt":188},{"_creationTime":305,"_id":306,"community":307,"display":308,"identity":317,"providers":318,"relations":321,"workflow":322},1778054691785.2563,"k175a63p5vhhc6160by5rqytgd867xs7",{"reviewCount":8},{"description":309,"installMethods":310,"name":311,"sourceUrl":312,"tags":313},"Implement AI image generation capabilities using the z-ai-web-dev-sdk. Use this skill when the user needs to create images from text descriptions, generate visual content, create artwork, design assets, or build applications with AI-powered image creation. Supports multiple image sizes and returns base64 encoded images. Also includes CLI tool for quick image generation.",{},"AI Image Generation","https://github.com/answerzhao/agent-skills/tree/HEAD/glm-skills/image-generation",[314,315,18,297,316,19],"image-generation","ai","typescript",{"githubOwner":190,"githubRepo":191,"locale":23,"slug":314,"type":192},{"extract":319,"llm":320},{"commitSha":197,"license":185},{"promptVersionExtension":169,"promptVersionScoring":170,"score":172,"targetMarket":175,"tier":176},{"repoId":200},{"anyEnrichmentAt":220,"extractAt":221,"githubAt":220,"llmAt":188,"updatedAt":188},{"_creationTime":324,"_id":325,"community":326,"display":327,"identity":337,"providers":340,"relations":345,"workflow":347},1778054035325.8745,"k178gqcwez20sdn3tt9v878bf5866y07",{"reviewCount":8},{"description":328,"installMethods":329,"name":330,"sourceUrl":331,"tags":332},"Generate AI voiceovers, sound effects, and music using ElevenLabs APIs. Use when creating audio content for videos, podcasts, or games. Triggers include generating voiceovers, narration, dialogue, sound effects from descriptions, background music, soundtrack generation, voice cloning, or any audio synthesis task.",{},"ElevenLabs Audio Generation","https://github.com/digitalsamba/claude-code-video-toolkit/tree/HEAD/.claude/skills/elevenlabs",[17,333,15,237,334,335,336],"voiceover","remotion","video","python",{"githubOwner":338,"githubRepo":339,"locale":23,"slug":237,"type":192},"digitalsamba","claude-code-video-toolkit",{"extract":341,"llm":343},{"commitSha":342,"license":185},"dc1bbd251ef137bde9cf460bacb88f13adb3a808",{"promptVersionExtension":169,"promptVersionScoring":170,"score":344,"targetMarket":175,"tier":176},93,{"repoId":346},"kd77w77a4w1f7nnb9v4fmh2eb1865dn1",{"anyEnrichmentAt":348,"extractAt":349,"githubAt":348,"llmAt":350,"updatedAt":350},1778054036248,1778054035325,1778054079849,{"_creationTime":352,"_id":353,"community":354,"display":355,"identity":364,"providers":366,"relations":372,"workflow":374},1778054061126.638,"k170rzj5b63rat15sg81fqcqyx867e52",{"reviewCount":8},{"description":356,"installMethods":357,"name":358,"sourceUrl":359,"tags":360},"Text-to-speech conversion using `uvx edge-tts` for generating audio from text. Use when (1) User requests audio/voice output with the \"tts\" trigger or keyword. (2) Content needs to be spoken rather than read (multitasking, accessibility, driving, cooking). (3) User wants a specific voice, speed, pitch, or format for TTS output.",{},"Edge TTS","https://github.com/aahl/skills/tree/HEAD/skills/edge-tts",[15,17,16,361,362,363],"uvx","edge-tts","microsoft",{"githubOwner":365,"githubRepo":272,"locale":23,"slug":362,"type":192},"aahl",{"extract":367,"llm":370},{"commitSha":368,"license":369},"503806b8502ad5965d31c46b9e46584f0746f33d","MIT-0",{"promptVersionExtension":169,"promptVersionScoring":170,"score":371,"targetMarket":175,"tier":176},85,{"repoId":373},"kd7f9kgmrb1hqjqtdjzws1v09d865znt",{"anyEnrichmentAt":375,"extractAt":376,"githubAt":375,"llmAt":377,"updatedAt":377},1778054061476,1778054061126,1778054102990]