[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"extension-skill-elevenlabs-voice-changer-zh-CN":3,"guides-for-elevenlabs-voice-changer":231,"similar-k17bjp4rs3rva38svbvgzcw685866c6c":232},{"_creationTime":4,"_id":5,"children":6,"community":7,"display":9,"evaluation":21,"identity":191,"isFallback":195,"parentExtension":196,"providers":197,"relations":201,"repo":203,"workflow":228},1778053440456.661,"k17bjp4rs3rva38svbvgzcw685866c6c",[],{"reviewCount":8},0,{"description":10,"installMethods":11,"name":12,"sourceUrl":13,"tags":14},"Transform the voice in an audio recording into a different target voice while preserving emotion, timing, and delivery using the ElevenLabs Voice Changer (speech-to-speech) API. Use when converting one voice to another, changing the speaker/narrator of an existing recording, dubbing a voice-over in a different voice, creating character voices from a scratch performance, anonymizing a speaker, or any \"voice conversion / voice transfer / speech-to-speech\" task. Make sure to use this skill whenever the user mentions voice changing, voice conversion, speech-to-speech, swapping a voice in audio, re-voicing a clip, or applying a different voice to an existing recording — even if they don't explicitly say \"voice changer\".",{},"ElevenLabs Voice Changer","https://github.com/elevenlabs/skills/tree/HEAD/voice-changer",[15,16,17,18,19,20],"audio","voice","elevenlabs","api","conversion","speech-to-speech",{"_creationTime":22,"_id":23,"extensionId":5,"locale":24,"result":25,"trustSignals":179,"workflow":189},1778053480675.1755,"kn7bbxvgmrstbet3g4zn65abvd86714f","en",{"checks":26,"evaluatedAt":169,"extensionSummary":170,"promptVersionExtension":171,"promptVersionScoring":172,"rationale":173,"score":174,"summary":175,"tags":176,"targetMarket":177,"tier":178},[27,32,35,38,42,45,49,53,57,60,64,69,72,76,79,82,85,88,91,94,98,102,106,110,114,117,120,123,128,131,134,137,140,143,147,150,153,156,159,162,166],{"category":28,"check":29,"severity":30,"summary":31},"Practical Utility","Problem relevance","pass","The description clearly states the user problem of transforming audio voice while preserving emotion, timing, and delivery, and provides specific use cases like dubbing and anonymization.",{"category":28,"check":33,"severity":30,"summary":34},"Unique selling proposition","The skill leverages the ElevenLabs Voice Changer API to perform speech-to-speech conversion, offering a distinct capability beyond standard text-to-speech or simple audio manipulation.",{"category":28,"check":36,"severity":30,"summary":37},"Production readiness","The skill provides clear setup instructions, parameter details, and code examples for multiple languages (Python, JavaScript, cURL), covering the end-to-end workflow for voice conversion.",{"category":39,"check":40,"severity":30,"summary":41},"Scope","Single responsibility principle","The extension's scope is focused solely on voice changing (speech-to-speech conversion) using the ElevenLabs API, aligning with its name and description.",{"category":39,"check":43,"severity":30,"summary":44},"Description quality","The description accurately and concisely explains the skill's functionality, use cases, and setup requirements, matching the provided code and documentation.",{"category":46,"check":47,"severity":30,"summary":48},"Invocation","Scoped tools","The skill acts as a wrapper around the ElevenLabs speech-to-speech API, which is a single, well-defined function, effectively acting as a single, narrow tool.",{"category":50,"check":51,"severity":30,"summary":52},"Documentation","Configuration & parameter reference","All parameters for the speech-to-speech conversion, including voice settings and output formats, are clearly documented with descriptions and defaults where applicable.",{"category":39,"check":54,"severity":55,"summary":56},"Tool naming","not_applicable","This extension functions as a single tool and does not expose multiple user-facing commands.",{"category":39,"check":58,"severity":30,"summary":59},"Minimal I/O surface","The input parameters are well-defined (voice_id, audio file, model_id, etc.), and the output is an audio stream, adhering to the principle of minimal I/O.",{"category":61,"check":62,"severity":30,"summary":63},"License","License usability","The extension is licensed under MIT, a permissive open-source license, with a LICENSE file present in the repository.",{"category":65,"check":66,"severity":67,"summary":68},"Maintenance","Commit recency","warning","There are no commits in the last 12 months on the default branch, suggesting the project may be unmaintained.",{"category":65,"check":70,"severity":55,"summary":71},"Dependency Management","No third-party dependencies are managed within this skill's repository itself; the skill interacts with an external API.",{"category":73,"check":74,"severity":30,"summary":75},"Security","Secret Management","The skill requires an ElevenLabs API key, which is handled via environment variables and not hardcoded or echoed in output, as indicated by the setup instructions and metadata.",{"category":73,"check":77,"severity":30,"summary":78},"Injection","The skill relies on a well-defined API and does not appear to load or execute untrusted third-party data as instructions.",{"category":73,"check":80,"severity":30,"summary":81},"Transitive Supply-Chain Grenades","The skill primarily interacts with the ElevenLabs API and does not appear to fetch or execute external code or content at runtime.",{"category":73,"check":83,"severity":30,"summary":84},"Sandbox Isolation","The skill's operation is confined to making API calls and does not involve modifying files outside its designated scope.",{"category":73,"check":86,"severity":30,"summary":87},"Sandbox escape primitives","No detached process spawns or retry loops around denied tool calls were found in the provided scripts or documentation.",{"category":73,"check":89,"severity":30,"summary":90},"Data Exfiltration","The skill's primary function is to call an external API with provided audio data; no mechanisms for unauthorized data exfiltration were detected.",{"category":73,"check":92,"severity":30,"summary":93},"Hidden Text Tricks","The bundled files appear to be free of hidden-steering tricks, control characters, or obfuscation intended to mislead the model.",{"category":95,"check":96,"severity":30,"summary":97},"Hooks","Opaque code execution","The provided code examples are clear and readable, with no signs of obfuscation like base64 payloads or minified code without source maps.",{"category":99,"check":100,"severity":30,"summary":101},"Portability","Structural Assumption","The skill assumes the presence of an ElevenLabs API key (via env var) and audio files, but these are standard requirements and not tied to specific user project structures.",{"category":103,"check":104,"severity":55,"summary":105},"Trust","Issues Attention","Issue tracking data is not available for this repository.",{"category":107,"check":108,"severity":67,"summary":109},"Versioning","Release Management","There is no explicit versioning information (e.g., version field in manifest, changelog, or release tags) available for this skill.",{"category":111,"check":112,"severity":30,"summary":113},"Code Execution","Validation","The skill interacts with a well-defined API, and the documentation outlines expected parameter types and formats, implying validation occurs at the API level.",{"category":73,"check":115,"severity":30,"summary":116},"Unguarded Destructive Operations","The skill's primary function is audio conversion, which is not a destructive operation.",{"category":111,"check":118,"severity":30,"summary":119},"Error Handling","The Python and cURL examples demonstrate basic error handling using try-except blocks and mention common API error codes.",{"category":111,"check":121,"severity":30,"summary":122},"Logging","The skill supports an `enable_logging` parameter for zero-retention mode, indicating a degree of control over logging, and Python/JS examples show basic stream handling which can be logged.",{"category":124,"check":125,"severity":126,"summary":127},"Compliance","GDPR","info","The skill processes audio data, which may contain personal data. While it sends this to the ElevenLabs API for processing, there is no indication of direct submission to a third party beyond the API provider itself, and the documentation does not specify explicit sanitization for personal data within the audio.",{"category":124,"check":129,"severity":30,"summary":130},"Target market","The skill is global in nature, focusing on audio processing without any explicit regional limitations or jurisdictional rules.",{"category":99,"check":132,"severity":30,"summary":133},"Runtime stability","The skill is designed to be cross-platform, relying on standard API interactions and providing examples for Python, JavaScript, and cURL, indicating good portability.",{"category":46,"check":135,"severity":30,"summary":136},"Precise Purpose","The description clearly defines the purpose (voice transformation) and use cases (dubbing, anonymization, character voices) and explicitly states what the skill does and when to use it.",{"category":46,"check":138,"severity":30,"summary":139},"Concise Frontmatter","The frontmatter in SKILL.md is concise and provides a clear summary of the core capability and trigger phrases within a reasonable character limit.",{"category":50,"check":141,"severity":30,"summary":142},"Concise Body","The SKILL.md file is well-structured, with the main body remaining concise and referencing external files for detailed installation instructions.",{"category":144,"check":145,"severity":30,"summary":146},"Context","Progressive Disclosure","The skill defers detailed installation instructions to a separate `references/installation.md` file, employing progressive disclosure.",{"category":144,"check":148,"severity":55,"summary":149},"Forked exploration","This skill is a direct API call and does not involve deep exploration or code review that would necessitate a forked context.",{"category":28,"check":151,"severity":30,"summary":152},"Usage examples","The skill provides ready-to-use code examples for Python, JavaScript, and cURL, demonstrating input, invocation, and expected output.",{"category":28,"check":154,"severity":30,"summary":155},"Edge cases","The documentation addresses edge cases like input length limits, file size limits, noisy audio, and provides guidance on input audio best practices and error handling.",{"category":111,"check":157,"severity":55,"summary":158},"Tool Fallback","This skill does not appear to rely on optional external tools like an MCP server; it directly calls the ElevenLabs API.",{"category":99,"check":160,"severity":30,"summary":161},"Stack assumptions","The skill clearly states its stack assumptions and prerequisites, such as requiring an ElevenLabs API key and providing examples for common runtimes like Python and Node.js.",{"category":163,"check":164,"severity":30,"summary":165},"Safety","Halt on unexpected state","The provided error handling examples suggest that the skill will report errors, and the API itself would likely halt on unexpected states or invalid inputs.",{"category":99,"check":167,"severity":30,"summary":168},"Cross-skill coupling","The skill is self-contained and focused on voice changing, with no apparent implicit reliance on other skills.",1778053463693,"This skill enables users to change the voice in an audio recording to a different target voice while maintaining the original performance characteristics. It supports various use cases including dubbing, character voice creation, and speaker anonymization, offering examples for Python, JavaScript, and cURL.","2.0.0","3.4.0","The 'voice-changer' skill is a well-documented and production-ready tool that effectively leverages the ElevenLabs API for speech-to-speech conversion. It provides clear setup instructions, multiple code examples, and handles common parameters and edge cases. Minor concerns include the lack of recent commits and explicit versioning, but these do not detract from its overall quality and usability.",95,"This skill transforms audio voice using the ElevenLabs API, preserving emotion and timing.",[15,16,17,18,19,20],"global","verified",{"codeQuality":180,"collectedAt":181,"documentation":182,"maintenance":184,"security":185,"testCoverage":188},{},1778053453169,{"descriptionLength":183,"readmeSize":8},724,{},{"hasNpmPackage":186,"license":187,"smitheryVerified":186},false,"MIT",{"hasCi":186,"hasTests":186},{"updatedAt":190},1778053480675,{"githubOwner":17,"githubRepo":192,"locale":24,"slug":193,"type":194},"skills","voice-changer","skill",true,null,{"extract":198,"llm":200},{"commitSha":199,"license":187},"b476f0ccf4be0e22b2e77cc39307665425d1472b",{"promptVersionExtension":171,"promptVersionScoring":172,"score":174,"targetMarket":177,"tier":178},{"repoId":202},"kd71z3hz1pg97d1k2d6kaqeqtx864knt",{"_creationTime":204,"_id":202,"identity":205,"providers":207,"workflow":225},1777995558409.8555,{"githubOwner":17,"githubRepo":192,"sourceUrl":206},"https://github.com/elevenlabs/skills",{"discover":208,"github":211},{"sources":209},[210],"skills-sh",{"closedIssues90d":212,"forks":213,"homepage":214,"license":187,"openIssues90d":215,"pushedAt":216,"readmeSize":217,"stars":218,"topics":219},16,22,"https://elevenlabs.io",1,1777909457000,3014,216,[220,17,221,222,192,223,224],"ai-agents","music","sfx","stt","tts",{"discoverAt":226,"extractAt":227,"githubAt":227,"updatedAt":227},1777995558409,1778053441433,{"anyEnrichmentAt":229,"extractAt":230,"githubAt":229,"llmAt":190,"updatedAt":190},1778053440833,1778053440456,[],[233,252,270,289,323,344],{"_creationTime":234,"_id":235,"community":236,"display":237,"identity":245,"providers":246,"relations":250,"workflow":251},1778053440456.6604,"k17a2cxtswmmk54b8wmpfbp5f9866jr0",{"reviewCount":8},{"description":238,"installMethods":239,"name":240,"sourceUrl":241,"tags":242},"Convert text to speech using ElevenLabs voice AI. Use when generating audio from text, creating voiceovers, building voice apps, or synthesizing speech in 70+ languages.",{},"ElevenLabs Text-to-Speech","https://github.com/elevenlabs/skills/tree/HEAD/text-to-speech",[243,17,16,15,244],"text-to-speech","synthesis",{"githubOwner":17,"githubRepo":192,"locale":24,"slug":243,"type":194},{"extract":247,"llm":248},{"commitSha":199,"license":187},{"promptVersionExtension":171,"promptVersionScoring":172,"score":249,"targetMarket":177,"tier":178},98,{"repoId":202},{"anyEnrichmentAt":229,"extractAt":230,"githubAt":229,"llmAt":190,"updatedAt":190},{"_creationTime":253,"_id":254,"community":255,"display":256,"identity":263,"providers":264,"relations":268,"workflow":269},1778053440456.658,"k17b8tkx3b4vgys5rp9avrjfmn866jdq",{"reviewCount":8},{"description":257,"installMethods":258,"name":259,"sourceUrl":260,"tags":261},"Generate music using ElevenLabs Music API. Use when creating instrumental tracks, songs with lyrics, background music, jingles, or any AI-generated music composition. Supports prompt-based generation, composition plans for granular control, and detailed output with metadata.",{},"ElevenLabs Music","https://github.com/elevenlabs/skills/tree/HEAD/music",[221,17,18,262,15],"generation",{"githubOwner":17,"githubRepo":192,"locale":24,"slug":221,"type":194},{"extract":265,"llm":266},{"commitSha":199,"license":187},{"promptVersionExtension":171,"promptVersionScoring":172,"score":267,"targetMarket":177,"tier":178},97,{"repoId":202},{"anyEnrichmentAt":229,"extractAt":230,"githubAt":229,"llmAt":190,"updatedAt":190},{"_creationTime":271,"_id":272,"community":273,"display":274,"identity":282,"providers":284,"relations":287,"workflow":288},1778053440456.6614,"k1721ymy7rr7zn8v2rpjnx1cvn86761x",{"reviewCount":8},{"description":275,"installMethods":276,"name":277,"sourceUrl":278,"tags":279},"Remove background noise and isolate vocals/speech from audio using ElevenLabs Voice Isolator (audio isolation) API. Use when cleaning up noisy recordings, removing music or background ambience from dialogue, isolating speech from field recordings, preparing audio for transcription, extracting vocals, or any \"denoise / clean up / isolate voice\" task.",{},"ElevenLabs Voice Isolator","https://github.com/elevenlabs/skills/tree/HEAD/voice-isolator",[15,16,17,18,280,281],"isolation","noise-reduction",{"githubOwner":17,"githubRepo":192,"locale":24,"slug":283,"type":194},"voice-isolator",{"extract":285,"llm":286},{"commitSha":199,"license":187},{"promptVersionExtension":171,"promptVersionScoring":172,"score":174,"targetMarket":177,"tier":178},{"repoId":202},{"anyEnrichmentAt":229,"extractAt":230,"githubAt":229,"llmAt":190,"updatedAt":190},{"_creationTime":290,"_id":291,"community":292,"display":293,"identity":307,"providers":311,"relations":316,"workflow":319},1778053197391.382,"k170wvt5rx3c1hv9a5sfkyezc1866k0q",{"reviewCount":8},{"description":294,"installMethods":295,"name":296,"sourceUrl":297,"tags":298},"Universal AI voice / text-to-speech skill supporting OpenAI TTS (gpt-4o-mini-tts, tts-1), ElevenLabs multilingual TTS with voice cloning, Bailian Qwen TTS (qwen-tts / qwen3-tts-vd with voice-design custom voices, long-text chunking built in), MiniMax speech-02-hd, SiliconFlow CosyVoice / SenseVoice, and PlayHT 2.0. Use this skill whenever the user asks to read text aloud, synthesize speech, generate narration, create voice-over, dub a script, or turn any text into audio (mp3 / wav / ogg / flac). Typical phrases include \"read this aloud\", \"generate voice for ...\", \"create a narration of ...\", \"tts this\", \"把这段念出来\", \"做个配音\", \"合成语音\", or mentions of voices / TTS model names like Alloy, Ash, Cherry, Rachel, CosyVoice, PlayHT. Always use this skill even if the user does not specify a provider — pick one from EXTEND.md defaults or available env keys.",{},"Happy Audio Gen","https://github.com/iamzhihuix/happy-claude-skills/tree/HEAD/skills/happy-audio-gen",[224,299,15,300,301,17,302,303,304,305,306],"speech","voice-generation","openai","bailian","minimax","siliconflow","playht","bun",{"githubOwner":308,"githubRepo":309,"locale":24,"slug":310,"type":194},"iamzhihuix","happy-claude-skills","happy-audio-gen",{"extract":312,"llm":314},{"commitSha":313,"license":187},"f49e7782a551759c9f9e0a4d4417ff053f0a86fd",{"promptVersionExtension":171,"promptVersionScoring":172,"score":315,"targetMarket":177,"tier":178},100,{"parentExtensionId":317,"repoId":318},"k173ydbbp6c0vdpxv5r0q9yvgd867en5","kd7dbbtdq95nkcs3k7fg9w6fdn864j0b",{"anyEnrichmentAt":320,"extractAt":321,"githubAt":320,"llmAt":322,"updatedAt":322},1778053199195,1778053197391,1778053284450,{"_creationTime":324,"_id":325,"community":326,"display":327,"identity":337,"providers":339,"relations":342,"workflow":343},1778053440456.6584,"k17120x7me8p1n30wxpg972esx866b8q",{"reviewCount":8},{"description":328,"installMethods":329,"name":330,"sourceUrl":331,"tags":332},"Transcribe audio to text using ElevenLabs Scribe. Supports batch transcription, realtime streaming from URLs, microphone input, and local files.",{},"ElevenLabs Speech-to-Text","https://github.com/elevenlabs/skills/tree/HEAD/openclaw/elevenlabs-transcribe",[333,15,17,334,335,336],"transcription","python","realtime","batch",{"githubOwner":17,"githubRepo":192,"locale":24,"slug":338,"type":194},"elevenlabs-transcribe",{"extract":340,"llm":341},{"commitSha":199,"license":187},{"promptVersionExtension":171,"promptVersionScoring":172,"score":249,"targetMarket":177,"tier":178},{"repoId":202},{"anyEnrichmentAt":229,"extractAt":230,"githubAt":229,"llmAt":190,"updatedAt":190},{"_creationTime":345,"_id":346,"community":347,"display":348,"identity":359,"providers":362,"relations":367,"workflow":369},1778054035325.875,"k1786xrb93cze519jccqw2h6hx867q00",{"reviewCount":8},{"description":349,"installMethods":350,"name":351,"sourceUrl":352,"tags":353},"Video and audio processing with FFmpeg. Use for format conversion, resizing, compression, audio extraction, and preparing assets for Remotion. Triggers include converting GIF to MP4, resizing video, extracting audio, compressing files, or any media transformation task.",{},"FFmpeg for Video Production","https://github.com/digitalsamba/claude-code-video-toolkit/tree/HEAD/.claude/skills/ffmpeg",[354,355,15,356,357,358,19],"ffmpeg","video","media","remotion","processing",{"githubOwner":360,"githubRepo":361,"locale":24,"slug":354,"type":194},"digitalsamba","claude-code-video-toolkit",{"extract":363,"llm":366},{"commitSha":364,"license":365},"dc1bbd251ef137bde9cf460bacb88f13adb3a808","MIT-0",{"promptVersionExtension":171,"promptVersionScoring":172,"score":174,"targetMarket":177,"tier":178},{"repoId":368},"kd77w77a4w1f7nnb9v4fmh2eb1865dn1",{"anyEnrichmentAt":370,"extractAt":371,"githubAt":370,"llmAt":372,"updatedAt":372},1778054036248,1778054035325,1778054079849]