[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"extension-skill-noizai-chat-with-anyone-hi":3,"guides-for-noizai-chat-with-anyone":221,"similar-k170fae6s3s4fwfbye7ys93sf5867w6j":222},{"_creationTime":4,"_id":5,"children":6,"community":7,"display":9,"evaluation":21,"identity":187,"isFallback":192,"parentExtension":193,"providers":194,"relations":198,"repo":200,"workflow":218},1778054291668.097,"k170fae6s3s4fwfbye7ys93sf5867w6j",[],{"reviewCount":8},0,{"description":10,"installMethods":11,"name":12,"sourceUrl":13,"tags":14},"Chat with any real person or fictional character in their own voice by automatically finding their speech online, extracting a clean reference sample, and generating audio replies. Also supports generating a matching voice from an uploaded image. Use when the user says \"我想跟xxx聊天\", \"你来扮演xxx跟我说话\", \"让xxx给我讲讲这篇文章\", \"我想跟图片中的人说话\", or similar.",{},"Chat with Anyone","https://github.com/noizai/skills/tree/HEAD/skills/chat-with-anyone",[15,16,17,18,19,20],"voice","tts","cloning","audio","ai","character",{"_creationTime":22,"_id":23,"extensionId":5,"locale":24,"result":25,"trustSignals":176,"workflow":185},1778054322837.3582,"kn7ams7c9mn0cvxbbj4khbar9x86785r","en",{"checks":26,"evaluatedAt":166,"extensionSummary":167,"promptVersionExtension":168,"promptVersionScoring":169,"rationale":170,"score":171,"summary":172,"tags":173,"targetMarket":174,"tier":175},[27,32,35,38,42,45,49,53,56,59,64,69,73,76,79,82,85,88,91,94,98,102,107,112,115,118,121,125,128,131,134,137,140,144,147,150,153,156,159,163],{"category":28,"check":29,"severity":30,"summary":31},"Practical Utility","Problem relevance","pass","The description clearly states the problem of chatting with characters in their own voice and provides specific trigger phrases.",{"category":28,"check":33,"severity":30,"summary":34},"Unique selling proposition","The skill uniquely addresses voice cloning from online sources and images, going beyond basic TTS or character impersonation.",{"category":28,"check":36,"severity":30,"summary":37},"Production readiness","The skill outlines clear workflows, prerequisites, and troubleshooting steps, indicating readiness for use in a production context.",{"category":39,"check":40,"severity":30,"summary":41},"Scope","Single responsibility principle","The skill focuses on generating voices and roleplaying, with clear workflows for name-based and image-based voice cloning, adhering to a single domain.",{"category":39,"check":43,"severity":30,"summary":44},"Description quality","The description accurately reflects the skill's functionality of finding speech online and generating audio replies, including image-based voice generation.",{"category":46,"check":47,"severity":30,"summary":48},"Invocation","Scoped tools","The skill uses specific Python scripts (`extract_ref_segment.py`, `voice_design.py`) and an external TTS skill, adhering to a verb-noun specialization.",{"category":50,"check":51,"severity":30,"summary":52},"Documentation","Configuration & parameter reference","Prerequisites like ffmpeg, yt-dlp, the tts skill, and NOIZ_API_KEY are clearly documented with verification and installation instructions.",{"category":39,"check":54,"severity":30,"summary":55},"Tool naming","Tool names like `extract_ref_segment.py` and `voice_design.py` are descriptive and follow kebab-case conventions.",{"category":39,"check":57,"severity":30,"summary":58},"Minimal I/O surface","Input parameters are clearly defined (paths, descriptions, keys) and outputs are specific files (WAV, voice_id.txt), avoiding unnecessary data dumps.",{"category":60,"check":61,"severity":62,"summary":63},"License","License usability","critical","No license file or SPDX identifier is present in the repository, making the licensing status unclear and potentially restrictive.",{"category":65,"check":66,"severity":67,"summary":68},"Maintenance","Commit recency","not_applicable","No commit history is available for this specific file/commit, so recency cannot be determined.",{"category":70,"check":71,"severity":30,"summary":72},"Security","Secret Management","The skill uses an environment variable (`NOIZ_API_KEY`) or a file (`~/.noiz_api_key`) for secrets, and the script `tts.py` has a config command to manage it, indicating appropriate handling.",{"category":70,"check":74,"severity":30,"summary":75},"Injection","The script's inputs are either file paths or command-line arguments that are processed by specific Python functions, with no indication of executing arbitrary external commands or data as instructions.",{"category":70,"check":77,"severity":30,"summary":78},"Transitive Supply-Chain Grenades","The skill relies on installed system binaries (ffmpeg, yt-dlp) and a separate skill (tts), but does not fetch or execute arbitrary code at runtime. All dependencies are documented.",{"category":70,"check":80,"severity":30,"summary":81},"Sandbox Isolation","The skill operates within its designated directory and temporary files, and interacts with external dependencies in a controlled manner, not modifying files outside its scope.",{"category":70,"check":83,"severity":30,"summary":84},"Sandbox escape primitives","No detached processes or retry loops around denied calls were found in the provided scripts.",{"category":70,"check":86,"severity":30,"summary":87},"Data Exfiltration","The skill requires an API key for an external service (Noiz) but handles it via environment variable or file, and does not appear to exfiltrate any additional confidential data.",{"category":70,"check":89,"severity":30,"summary":90},"Hidden Text Tricks","The bundled scripts and SKILL.md file do not contain any hidden text tricks, invisible characters, or obfuscated instructions.",{"category":70,"check":92,"severity":30,"summary":93},"Opaque code execution","The bundled Python scripts are standard, readable Python code, and do not contain obfuscation, base64 payloads, or eval calls.",{"category":95,"check":96,"severity":30,"summary":97},"Portability","Structural Assumption","The skill assumes the presence of system binaries and Python packages which are clearly documented, and uses relative paths for its own files.",{"category":99,"check":100,"severity":67,"summary":101},"Trust","Issues Attention","Issue data is not available for this repository.",{"category":103,"check":104,"severity":105,"summary":106},"Versioning","Release Management","warning","There is no version information (e.g., version field in manifest, CHANGELOG, or GitHub releases) for this skill.",{"category":108,"check":109,"severity":110,"summary":111},"Code Execution","Validation","info","Input validation appears to be handled by Python's standard argument parsing, but explicit schema validation libraries are not evident for all inputs like file paths.",{"category":70,"check":113,"severity":30,"summary":114},"Unguarded Destructive Operations","The skill's operations are primarily file manipulation and API calls, with no destructive operations like deletion or modification of user data outside its temporary scope.",{"category":108,"check":116,"severity":30,"summary":117},"Error Handling","Python scripts include basic error handling, explicit error messages, and exit codes, and the use of `try...except` blocks is implied for subprocess calls.",{"category":108,"check":119,"severity":30,"summary":120},"Logging","The skill prints status messages and errors to stdout/stderr, providing an audit trail of its execution for troubleshooting.",{"category":122,"check":123,"severity":110,"summary":124},"Compliance","GDPR","The skill processes audio and image data which could potentially contain personal data. While not directly submitting PII to a third party without consent, care should be taken with uploaded images and reference audio sources.",{"category":122,"check":126,"severity":30,"summary":127},"Target market","The skill's functionality is global and does not appear to be restricted to any specific geographic or legal jurisdiction.",{"category":95,"check":129,"severity":30,"summary":130},"Runtime stability","The skill relies on common system binaries (ffmpeg, yt-dlp) and Python, which are generally available across platforms, with clear prerequisites documented.",{"category":46,"check":132,"severity":30,"summary":133},"Precise Purpose","The skill's description and SKILL.md clearly define its purpose: cloning voices from online speech or images, and provides specific use cases and ethical guidelines.",{"category":46,"check":135,"severity":30,"summary":136},"Concise Frontmatter","The frontmatter in SKILL.md is concise, clearly stating the skill's name, description, permissions, and essential metadata like the API key requirement.",{"category":50,"check":138,"severity":30,"summary":139},"Concise Body","The SKILL.md body is well-structured with clear sections for workflows, prerequisites, and examples, keeping the main content focused.",{"category":141,"check":142,"severity":30,"summary":143},"Context","Progressive Disclosure","Detailed workflows, prerequisites, and examples are provided within SKILL.md, with no large embedded blobs or complex nested structures requiring external files.",{"category":141,"check":145,"severity":67,"summary":146},"Forked exploration","The skill is not designed for deep exploration that would require forking context.",{"category":28,"check":148,"severity":30,"summary":149},"Usage examples","The SKILL.md includes two detailed, end-to-end examples (name-based and image-based) showing user input, agent steps, and expected outcomes.",{"category":28,"check":151,"severity":30,"summary":152},"Edge cases","The skill documentation addresses potential issues like download failures, missing subtitles, voice design errors, and offers troubleshooting steps.",{"category":108,"check":154,"severity":110,"summary":155},"Tool Fallback","While the skill lists `tts` as a dependency, it doesn't explicitly state if it's optional or if there's a fallback if the `tts` skill is unavailable, though it implies its use.",{"category":95,"check":157,"severity":30,"summary":158},"Stack assumptions","The skill clearly declares its system dependencies (ffmpeg, yt-dlp) and Python requirements, along with the need for the `tts` skill.",{"category":160,"check":161,"severity":30,"summary":162},"Safety","Halt on unexpected state","The skill's scripts and documentation imply that errors will halt execution, and troubleshooting steps are provided for common failures.",{"category":95,"check":164,"severity":110,"summary":165},"Cross-skill coupling","The skill depends on the `tts` skill, which is mentioned in prerequisites but not explicitly handled as an optional dependency with a fallback within this skill's documentation.",1778054305475,"This skill enables users to generate audio replies in the voice of real or fictional characters by cloning speech from online videos or from uploaded images. It leverages external tools like ffmpeg, yt-dlp, and a separate TTS skill, requiring specific API keys for advanced features.","2.0.0","3.4.0","The extension is well-documented, provides clear workflows and examples, and handles its dependencies and security considerations appropriately. The main drawback is the lack of a clear license, which is a critical issue, and the absence of versioning information.",85,"A robust skill for cloning voices from online media or images to generate audio replies, with detailed documentation and ethical considerations.",[15,16,17,18,19,20],"global","flagged",{"codeQuality":177,"collectedAt":178,"documentation":179,"maintenance":181,"security":182,"testCoverage":184},{},1778054293440,{"descriptionLength":180,"readmeSize":8},338,{},{"hasNpmPackage":183,"smitheryVerified":183},false,{"hasCi":183,"hasTests":183},{"updatedAt":186},1778054322837,{"githubOwner":188,"githubRepo":189,"locale":24,"slug":190,"type":191},"noizai","skills","chat-with-anyone","skill",true,null,{"extract":195,"llm":197},{"commitSha":196,"license":62},"3428725f8c34e2179e0bdaf4c441697e0014d289",{"promptVersionExtension":168,"promptVersionScoring":169,"score":171,"targetMarket":174,"tier":175},{"repoId":199},"kd76a0e9dvgjrpvm0qe8k24qy18653ah",{"_creationTime":201,"_id":199,"identity":202,"providers":204,"workflow":215},1777995558409.862,{"githubOwner":188,"githubRepo":189,"sourceUrl":203},"https://github.com/noizai/skills",{"discover":205,"github":208},{"sources":206},[207],"skills-sh",{"closedIssues90d":8,"forks":209,"openIssues90d":210,"pushedAt":211,"readmeSize":212,"stars":213,"topics":214},69,2,1776527927000,4484,490,[],{"discoverAt":216,"extractAt":217,"githubAt":217,"updatedAt":217},1777995558409,1778054292661,{"anyEnrichmentAt":219,"extractAt":220,"githubAt":219,"llmAt":186,"updatedAt":186},1778054291966,1778054291668,[],[223,260,281,304,325,344],{"_creationTime":224,"_id":225,"community":226,"display":227,"identity":242,"providers":246,"relations":253,"workflow":256},1778053197391.382,"k170wvt5rx3c1hv9a5sfkyezc1866k0q",{"reviewCount":8},{"description":228,"installMethods":229,"name":230,"sourceUrl":231,"tags":232},"Universal AI voice / text-to-speech skill supporting OpenAI TTS (gpt-4o-mini-tts, tts-1), ElevenLabs multilingual TTS with voice cloning, Bailian Qwen TTS (qwen-tts / qwen3-tts-vd with voice-design custom voices, long-text chunking built in), MiniMax speech-02-hd, SiliconFlow CosyVoice / SenseVoice, and PlayHT 2.0. Use this skill whenever the user asks to read text aloud, synthesize speech, generate narration, create voice-over, dub a script, or turn any text into audio (mp3 / wav / ogg / flac). Typical phrases include \"read this aloud\", \"generate voice for ...\", \"create a narration of ...\", \"tts this\", \"把这段念出来\", \"做个配音\", \"合成语音\", or mentions of voices / TTS model names like Alloy, Ash, Cherry, Rachel, CosyVoice, PlayHT. Always use this skill even if the user does not specify a provider — pick one from EXTEND.md defaults or available env keys.",{},"Happy Audio Gen","https://github.com/iamzhihuix/happy-claude-skills/tree/HEAD/skills/happy-audio-gen",[16,233,18,234,235,236,237,238,239,240,241],"speech","voice-generation","openai","elevenlabs","bailian","minimax","siliconflow","playht","bun",{"githubOwner":243,"githubRepo":244,"locale":24,"slug":245,"type":191},"iamzhihuix","happy-claude-skills","happy-audio-gen",{"extract":247,"llm":250},{"commitSha":248,"license":249},"f49e7782a551759c9f9e0a4d4417ff053f0a86fd","MIT",{"promptVersionExtension":168,"promptVersionScoring":169,"score":251,"targetMarket":174,"tier":252},100,"verified",{"parentExtensionId":254,"repoId":255},"k173ydbbp6c0vdpxv5r0q9yvgd867en5","kd7dbbtdq95nkcs3k7fg9w6fdn864j0b",{"anyEnrichmentAt":257,"extractAt":258,"githubAt":257,"llmAt":259,"updatedAt":259},1778053199195,1778053197391,1778053284450,{"_creationTime":261,"_id":262,"community":263,"display":264,"identity":273,"providers":275,"relations":279,"workflow":280},1778054291668.0964,"k17477k0d5jgrwmqqkgp1phgen866eck",{"reviewCount":8},{"description":265,"installMethods":266,"name":267,"sourceUrl":268,"tags":269},"Use this skill whenever the user wants speech to sound more human, companion-like, or emotionally expressive. Triggers include: any mention of 'say like', 'talk like', 'speak like', 'companion voice', 'comfort me', 'cheer me up', 'sound more human', 'good night voice', 'good morning voice', or requests to add fillers, emotion, or personality to generated speech. Also use when the user wants to mimic a specific character's voice, apply speaking style presets (goodnight, morning, comfort, celebration, chatting), tune emotional parameters like warmth or tenderness, or make TTS output feel like a real person talking. If the user asks for a 'voice message', 'companion audio', 'character voice', or wants speech that sighs, laughs, hesitates, or sounds genuinely warm, use this skill. Do NOT use for plain text-to-speech without personality, music generation, sound effects, or general coding tasks unrelated to expressive speech.",{},"Characteristic Voice","https://github.com/noizai/skills/tree/HEAD/skills/characteristic-voice",[16,233,15,270,271,272],"emotion","companion","script",{"githubOwner":188,"githubRepo":189,"locale":24,"slug":274,"type":191},"characteristic-voice",{"extract":276,"llm":277},{"commitSha":196,"license":67},{"promptVersionExtension":168,"promptVersionScoring":169,"score":278,"targetMarket":174,"tier":252},98,{"repoId":199},{"anyEnrichmentAt":219,"extractAt":220,"githubAt":219,"llmAt":186,"updatedAt":186},{"_creationTime":282,"_id":283,"community":284,"display":285,"identity":293,"providers":294,"relations":298,"workflow":300},1778053440456.6604,"k17a2cxtswmmk54b8wmpfbp5f9866jr0",{"reviewCount":8},{"description":286,"installMethods":287,"name":288,"sourceUrl":289,"tags":290},"Convert text to speech using ElevenLabs voice AI. Use when generating audio from text, creating voiceovers, building voice apps, or synthesizing speech in 70+ languages.",{},"ElevenLabs Text-to-Speech","https://github.com/elevenlabs/skills/tree/HEAD/text-to-speech",[291,236,15,18,292],"text-to-speech","synthesis",{"githubOwner":236,"githubRepo":189,"locale":24,"slug":291,"type":191},{"extract":295,"llm":297},{"commitSha":296,"license":249},"b476f0ccf4be0e22b2e77cc39307665425d1472b",{"promptVersionExtension":168,"promptVersionScoring":169,"score":278,"targetMarket":174,"tier":252},{"repoId":299},"kd71z3hz1pg97d1k2d6kaqeqtx864knt",{"anyEnrichmentAt":301,"extractAt":302,"githubAt":301,"llmAt":303,"updatedAt":303},1778053440833,1778053440456,1778053480675,{"_creationTime":305,"_id":306,"community":307,"display":308,"identity":317,"providers":319,"relations":323,"workflow":324},1778053440456.6614,"k1721ymy7rr7zn8v2rpjnx1cvn86761x",{"reviewCount":8},{"description":309,"installMethods":310,"name":311,"sourceUrl":312,"tags":313},"Remove background noise and isolate vocals/speech from audio using ElevenLabs Voice Isolator (audio isolation) API. Use when cleaning up noisy recordings, removing music or background ambience from dialogue, isolating speech from field recordings, preparing audio for transcription, extracting vocals, or any \"denoise / clean up / isolate voice\" task.",{},"ElevenLabs Voice Isolator","https://github.com/elevenlabs/skills/tree/HEAD/voice-isolator",[18,15,236,314,315,316],"api","isolation","noise-reduction",{"githubOwner":236,"githubRepo":189,"locale":24,"slug":318,"type":191},"voice-isolator",{"extract":320,"llm":321},{"commitSha":296,"license":249},{"promptVersionExtension":168,"promptVersionScoring":169,"score":322,"targetMarket":174,"tier":252},95,{"repoId":299},{"anyEnrichmentAt":301,"extractAt":302,"githubAt":301,"llmAt":303,"updatedAt":303},{"_creationTime":326,"_id":327,"community":328,"display":329,"identity":337,"providers":339,"relations":342,"workflow":343},1778053440456.661,"k17bjp4rs3rva38svbvgzcw685866c6c",{"reviewCount":8},{"description":330,"installMethods":331,"name":332,"sourceUrl":333,"tags":334},"Transform the voice in an audio recording into a different target voice while preserving emotion, timing, and delivery using the ElevenLabs Voice Changer (speech-to-speech) API. Use when converting one voice to another, changing the speaker/narrator of an existing recording, dubbing a voice-over in a different voice, creating character voices from a scratch performance, anonymizing a speaker, or any \"voice conversion / voice transfer / speech-to-speech\" task. Make sure to use this skill whenever the user mentions voice changing, voice conversion, speech-to-speech, swapping a voice in audio, re-voicing a clip, or applying a different voice to an existing recording — even if they don't explicitly say \"voice changer\".",{},"ElevenLabs Voice Changer","https://github.com/elevenlabs/skills/tree/HEAD/voice-changer",[18,15,236,314,335,336],"conversion","speech-to-speech",{"githubOwner":236,"githubRepo":189,"locale":24,"slug":338,"type":191},"voice-changer",{"extract":340,"llm":341},{"commitSha":296,"license":249},{"promptVersionExtension":168,"promptVersionScoring":169,"score":322,"targetMarket":174,"tier":252},{"repoId":299},{"anyEnrichmentAt":301,"extractAt":302,"githubAt":301,"llmAt":303,"updatedAt":303},{"_creationTime":345,"_id":346,"community":347,"display":348,"identity":356,"providers":359,"relations":363,"workflow":365},1778054691785.2524,"k1712xyy3wyvy83c0f9z7kccg9866jg4",{"reviewCount":8},{"description":349,"installMethods":350,"name":351,"sourceUrl":352,"tags":353},"Implement text-to-speech (TTS) capabilities using the z-ai-web-dev-sdk. Use this skill when the user needs to convert text into natural-sounding speech, create audio content, build voice-enabled applications, or generate spoken audio files. Supports multiple voices, adjustable speed, and various audio formats.",{},"Text-to-Speech (TTS)","https://github.com/answerzhao/agent-skills/tree/HEAD/glm-skills/TTS",[16,291,18,354,355],"sdk","z-ai-web-dev-sdk",{"githubOwner":357,"githubRepo":358,"locale":24,"slug":16,"type":191},"answerzhao","agent-skills",{"extract":360,"llm":362},{"commitSha":361,"license":249},"aad73edbd0d9ffbc3d6a402b6eafa6dab96d5ebb",{"promptVersionExtension":168,"promptVersionScoring":169,"score":322,"targetMarket":174,"tier":252},{"repoId":364},"kd712v2g1pay70swwj0jpv2ggs864zgh",{"anyEnrichmentAt":366,"extractAt":367,"githubAt":366,"llmAt":368,"updatedAt":368},1778054692243,1778054691785,1778054738050]