[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"extension-skill-NoizAI-characteristic-voice-en":3,"guides-for-NoizAI-characteristic-voice":374,"similar-k170p7rk3v2rqw6qxbn248pbt986nhzb-en":375},{"_creationTime":4,"_id":5,"children":6,"community":7,"display":9,"evaluation":15,"identity":247,"isFallback":232,"parentExtension":252,"providers":253,"relations":258,"repo":260,"tags":370,"workflow":371},1778694659798.2534,"k170p7rk3v2rqw6qxbn248pbt986nhzb",[],{"reviewCount":8},0,{"description":10,"installMethods":11,"name":13,"sourceUrl":14},"Use this skill whenever the user wants speech to sound more human, companion-like, or emotionally expressive. Triggers include: any mention of 'say like', 'talk like', 'speak like', 'companion voice', 'comfort me', 'cheer me up', 'sound more human', 'good night voice', 'good morning voice', or requests to add fillers, emotion, or personality to generated speech. Also use when the user wants to mimic a specific character's voice, apply speaking style presets (goodnight, morning, comfort, celebration, chatting), tune emotional parameters like warmth or tenderness, or make TTS output feel like a real person talking. If the user asks for a 'voice message', 'companion audio', 'character voice', or wants speech that sighs, laughs, hesitates, or sounds genuinely warm, use this skill. Do NOT use for plain text-to-speech without personality, music generation, sound effects, or general coding tasks unrelated to expressive speech.",{"claudeCode":12},"NoizAI/skills","characteristic-voice","https://github.com/NoizAI/skills",{"_creationTime":16,"_id":17,"extensionId":5,"locale":18,"result":19,"trustSignals":230,"workflow":245},1778694682311.9746,"kn7d40m62q7fzgdd0zpkrb4fq586nwvz","en",{"checks":20,"evaluatedAt":197,"extensionSummary":198,"features":199,"nonGoals":205,"promptVersionExtension":210,"promptVersionScoring":211,"purpose":212,"rationale":213,"score":214,"summary":215,"tags":216,"targetMarket":223,"tier":224,"useCases":225},[21,26,29,32,36,39,44,48,51,54,58,62,66,70,73,76,79,82,85,88,92,96,100,105,109,112,115,118,122,125,128,131,134,137,140,144,148,152,155,159,162,165,168,171,175,178,181,184,187,190,194],{"category":22,"check":23,"severity":24,"summary":25},"Practical Utility","Problem relevance","pass","The description clearly states the problem of making AI speech sound more human, companion-like, and emotionally expressive, with specific triggers.",{"category":22,"check":27,"severity":24,"summary":28},"Unique selling proposition","The skill offers significant value over simple prompting by providing specialized presets, emotion tuning, and the ability to use reference audio for voice cloning, which goes beyond default LLM capabilities.",{"category":22,"check":30,"severity":24,"summary":31},"Production readiness","The skill is production-ready, providing a complete workflow from configuration to audio output, with clear instructions for prerequisites and backend choices.",{"category":33,"check":34,"severity":24,"summary":35},"Scope","Single responsibility principle","The skill focuses on enhancing speech characteristics and is clearly scoped to TTS manipulation, without venturing into unrelated domains.",{"category":33,"check":37,"severity":24,"summary":38},"Description quality","The displayed description accurately and concisely reflects the skill's capabilities, including specific triggers, presets, and non-goals.",{"category":40,"check":41,"severity":42,"summary":43},"Invocation","Scoped tools","not_applicable","This skill is a wrapper script and does not expose individual tools to the agent.",{"category":45,"check":46,"severity":24,"summary":47},"Documentation","Configuration & parameter reference","All options and parameters for the `speak.sh` script are documented in the help text, including presets, emotion tuning, and backend options.",{"category":33,"check":49,"severity":42,"summary":50},"Tool naming","The skill uses a single script (`speak.sh`) and a configuration command, without exposing multiple distinct tools.",{"category":33,"check":52,"severity":42,"summary":53},"Minimal I/O surface","This is a script-based skill, not tool-based, so this check is not applicable.",{"category":55,"check":56,"severity":24,"summary":57},"License","License usability","The project appears to be under a permissive open-source license, indicated by common OSS practices and lack of restrictive statements, although a formal LICENSE file is not directly provided in the excerpt. The `README.md` does not suggest non-permissive licensing.",{"category":59,"check":60,"severity":24,"summary":61},"Maintenance","Commit recency","The last commit was on May 7, 2026, which is within the last 3 months.",{"category":59,"check":63,"severity":64,"summary":65},"Dependency Management","warning","The skill lists Python, curl, and potentially others like `kokoro-tts`, `yt-dlp`, `ffmpeg`, `rg` as prerequisites, but does not provide a mechanism for managing or updating these dependencies automatically, relying on manual user installation.",{"category":67,"check":68,"severity":24,"summary":69},"Security","Secret Management","The API key is handled securely by saving it to `~/.noiz_api_key` with mode 600 permissions and is normalized to base64. It's not directly echoed or hardcoded in the script.",{"category":67,"check":71,"severity":24,"summary":72},"Injection","The script primarily processes user-provided text and audio files. There are no indications of loading external code or treating fetched data as executable instructions.",{"category":67,"check":74,"severity":24,"summary":75},"Transitive Supply-Chain Grenades","The script relies on bundled prerequisites and local execution of `kokoro-tts` or API calls to Noiz. It does not fetch remote scripts or markdown to execute.",{"category":67,"check":77,"severity":24,"summary":78},"Sandbox Isolation","The script interacts with user-provided text and audio, and writes output to a specified file. It handles API keys securely and does not modify files outside its intended scope.",{"category":67,"check":80,"severity":24,"summary":81},"Sandbox escape primitives","The script does not exhibit detached-process spawns or retry loops around denied tool calls.",{"category":67,"check":83,"severity":24,"summary":84},"Data Exfiltration","Data is sent to the Noiz API only when explicitly chosen by the user and is documented. The Kokoro backend runs locally. No confidential data is exfiltrated by default.",{"category":67,"check":86,"severity":24,"summary":87},"Hidden Text Tricks","The bundled scripts and documentation do not contain hidden-steering tricks, invisible characters, or other obfuscation methods.",{"category":89,"check":90,"severity":24,"summary":91},"Hooks","Opaque code execution","The scripts are written in plain bash and python, with no signs of obfuscation, base64 payloads, or runtime script fetching.",{"category":93,"check":94,"severity":24,"summary":95},"Portability","Structural Assumption","The script assumes the presence of common tools like curl and python3, and saves its API key to a standard user-specific location, which are reasonable assumptions.",{"category":97,"check":98,"severity":24,"summary":99},"Trust","Issues Attention","There are 2 issues opened in the last 90 days and 0 closed, indicating active but early issue engagement.",{"category":101,"check":102,"severity":103,"summary":104},"Versioning","Release Management","info","The `SKILL.md` frontmatter does not explicitly declare a version, and installation instructions refer to `main` (via `NoizAI/skills --full-depth`), but the commit recency check passes, indicating active development.",{"category":106,"check":107,"severity":24,"summary":108},"Code Execution","Validation","The `speak.sh` script has argument parsing that validates input types and values, such as checking for required arguments and valid presets.",{"category":67,"check":110,"severity":24,"summary":111},"Unguarded Destructive Operations","The script's primary function is audio generation, which is not a destructive operation. It writes to an output file, which is a standard I/O operation.",{"category":106,"check":113,"severity":24,"summary":114},"Error Handling","The script includes error handling for missing arguments, invalid presets, API key configuration, and backend availability, exiting with informative messages.",{"category":106,"check":116,"severity":42,"summary":117},"Logging","The script provides informative output to stderr regarding its operations but does not write to a structured audit log file.",{"category":119,"check":120,"severity":24,"summary":121},"Compliance","GDPR","The skill handles personal data (user text, reference audio) by processing it either locally or sending it to the Noiz API, with clear documentation on data transmission and privacy. Reference audio is only uploaded if explicitly provided.",{"category":119,"check":123,"severity":24,"summary":124},"Target market","The skill is globally applicable and does not contain any region-specific logic or language constraints.",{"category":93,"check":126,"severity":24,"summary":127},"Runtime stability","The script relies on common tools like curl and python3 and is designed for POSIX-compliant shells, ensuring broad runtime stability.",{"category":45,"check":129,"severity":24,"summary":130},"README","The README file exists and provides a good overview of the repository's skills, including installation instructions and highlights.",{"category":33,"check":132,"severity":42,"summary":133},"Tool surface size","This skill is a script wrapper with a configuration command, not a collection of multiple tools.",{"category":40,"check":135,"severity":42,"summary":136},"Overlapping near-synonym tools","The skill does not expose multiple tools that could be considered near-synonyms.",{"category":45,"check":138,"severity":24,"summary":139},"Phantom features","All advertised features, such as presets and backend options, are implemented and documented in the `speak.sh` script.",{"category":141,"check":142,"severity":24,"summary":143},"Install","Installation instruction","The README provides clear installation instructions using `npx skills add` and the `speak.sh` script includes installation hints and a `config` command for API key setup.",{"category":145,"check":146,"severity":24,"summary":147},"Errors","Actionable error messages","Errors related to missing arguments, configuration, or backend issues provide clear explanations and suggest remediation steps.",{"category":149,"check":150,"severity":64,"summary":151},"Execution","Pinned dependencies","While common tools are assumed, specific external dependencies like `kokoro-tts`, `yt-dlp`, etc., are not pinned or managed by the skill, potentially leading to compatibility issues.",{"category":33,"check":153,"severity":42,"summary":154},"Dry-run preview","The skill's core function is audio generation, which is not a destructive operation and therefore does not require a dry-run mode.",{"category":156,"check":157,"severity":42,"summary":158},"Protocol","Idempotent retry & timeouts","The skill's operations are primarily local file writes or API calls to a third-party service; idempotency and timeouts are handled by the underlying TTS service or are not applicable to the script's direct operations.",{"category":119,"check":160,"severity":24,"summary":161},"Telemetry opt-in","The skill does not appear to emit telemetry by default. Any data transmission to the Noiz API requires explicit user action (providing text/audio and selecting the backend).",{"category":40,"check":163,"severity":24,"summary":164},"Precise Purpose","The description clearly states the skill's purpose is to make speech sound more human and companion-like, detailing specific triggers and use cases, and explicitly listing non-goals.",{"category":40,"check":166,"severity":24,"summary":167},"Concise Frontmatter","The `SKILL.md` frontmatter is concise, clearly stating the core purpose and listing trigger phrases.",{"category":45,"check":169,"severity":24,"summary":170},"Concise Body","The `SKILL.md` is concise, detailing prerequisites, privacy, presets, and usage without excessive length.",{"category":172,"check":173,"severity":42,"summary":174},"Context","Progressive Disclosure","The skill is relatively simple and does not require complex procedures that would benefit from progressive disclosure into separate reference files.",{"category":172,"check":176,"severity":42,"summary":177},"Forked exploration","This skill is not designed for deep exploration or code review; it performs a single, well-defined task.",{"category":22,"check":179,"severity":24,"summary":180},"Usage examples","The README and script help provide clear, ready-to-use examples for configuring the API key and generating speech with presets.",{"category":22,"check":182,"severity":24,"summary":183},"Edge cases","The script handles edge cases such as missing required arguments, invalid presets, and configuration issues with informative error messages.",{"category":106,"check":185,"severity":24,"summary":186},"Tool Fallback","The skill supports both a remote Noiz backend (requiring an API key) and a local Kokoro backend, allowing users to choose based on their needs and environment.",{"category":93,"check":188,"severity":24,"summary":189},"Stack assumptions","The script declares its runtime requirements (curl, python3) and recommends installation of specific tools like kokoro-tts if needed, making stack assumptions explicit.",{"category":191,"check":192,"severity":24,"summary":193},"Safety","Halt on unexpected state","The script checks for necessary preconditions like API keys and available backends, and halts execution with informative errors if they are not met.",{"category":93,"check":195,"severity":42,"summary":196},"Cross-skill coupling","This skill operates independently and does not appear to implicitly rely on or interact with other specific skills in the repository.",1778694682013,"This skill provides a command-line script (`speak.sh`) to enhance text-to-speech output, making it sound more human, companion-like, or emotionally expressive. It supports presets, custom emotion tuning, filler sounds, and voice cloning via reference audio using either a local backend (Kokoro) or a remote API (Noiz).",[200,201,202,203,204],"Make speech sound human, companion-like, or emotionally expressive","Incorporate non-lexical fillers (hmm, haha, aww)","Tune emotional parameters (warmth, tenderness, joy)","Use speaking style presets (goodnight, morning, comfort)","Mimic character voices using reference audio",[206,207,208,209],"Plain text-to-speech without added personality or emotion","Music generation or sound effect creation","General coding tasks unrelated to expressive speech generation","Directly controlling the underlying TTS engine beyond presets and parameters","3.0.0","4.4.0","To enable AI agents to communicate with a more natural, human-like, and emotionally resonant voice, fostering a greater sense of companionship and connection.","The skill is well-documented, secure, and offers a clear value proposition with good usability. The only minor deductions relate to dependency management and versioning clarity.",95,"A high-quality skill that adds emotional expressiveness and personality to AI-generated speech.",[217,218,219,220,221,222],"tts","voice","speech","emotion","character","nlp","global","community",[226,227,228,229],"Creating voice messages that feel genuinely warm and personal","Developing AI companions that exhibit personality and emotional depth","Adding subtle humanizing elements like sighs or hesitations to TTS","Generating audio with specific emotional tones for storytelling or role-playing",{"codeQuality":231,"collectedAt":233,"documentation":234,"maintenance":237,"security":242,"testCoverage":243},{"hasLockfile":232},false,1778694663695,{"descriptionLength":235,"readmeSize":236},933,4484,{"closedIssues90d":8,"forks":238,"hasChangelog":232,"openIssues90d":239,"pushedAt":240,"stars":241},72,2,1778143310000,497,{"hasNpmPackage":232,"smitheryVerified":232},{"hasCi":244,"hasTests":244},true,{"updatedAt":246},1778694682312,{"basePath":248,"githubOwner":249,"githubRepo":250,"locale":18,"slug":13,"type":251},"skills/characteristic-voice","NoizAI","skills","skill",null,{"evaluate":254,"extract":256},{"promptVersionExtension":210,"promptVersionScoring":211,"score":214,"tags":255,"targetMarket":223,"tier":224},[217,218,219,220,221,222],{"commitSha":257},"HEAD",{"repoId":259},"kd76ab29g8nkd2m33vabs03ybx86my58",{"_creationTime":261,"_id":259,"identity":262,"providers":263,"workflow":366},1778694655410.099,{"githubOwner":249,"githubRepo":250,"sourceUrl":14},{"classify":264,"discover":360,"github":363},{"commitSha":257,"extensions":265},[266,277,289,297,310,320,330,348],{"basePath":248,"description":10,"displayName":13,"installMethods":267,"rationale":268,"selectedPaths":269,"source":276,"sourceLanguage":18,"type":251},{"claudeCode":12},"SKILL.md frontmatter at skills/characteristic-voice/SKILL.md",[270,273],{"path":271,"priority":272},"SKILL.md","mandatory",{"path":274,"priority":275},"scripts/speak.sh","low","rule",{"basePath":278,"description":279,"displayName":280,"installMethods":281,"rationale":282,"selectedPaths":283,"source":276,"sourceLanguage":18,"type":251},"skills/chat-with-anyone","Chat with any real person or fictional character in their own voice by automatically finding their speech online, extracting a clean reference sample, and generating audio replies. Also supports generating a matching voice from an uploaded image. Use when the user says \"我想跟xxx聊天\", \"你来扮演xxx跟我说话\", \"让xxx给我讲讲这篇文章\", \"我想跟图片中的人说话\", or similar.","chat-with-anyone",{"claudeCode":12},"SKILL.md frontmatter at skills/chat-with-anyone/SKILL.md",[284,285,287],{"path":271,"priority":272},{"path":286,"priority":275},"scripts/extract_ref_segment.py",{"path":288,"priority":275},"scripts/voice_design.py",{"basePath":290,"description":291,"displayName":292,"installMethods":293,"rationale":294,"selectedPaths":295,"source":276,"sourceLanguage":18,"type":251},"skills/daily-news-caster","Fetches the latest news using news-aggregator-skill, formats it into a podcast script in Markdown format, and uses the tts skill to generate a podcast audio file. Use when the user asks to get the latest news and read it out as a podcast.","daily-news-caster",{"claudeCode":12},"SKILL.md frontmatter at skills/daily-news-caster/SKILL.md",[296],{"path":271,"priority":272},{"basePath":298,"description":299,"displayName":300,"installMethods":301,"rationale":302,"selectedPaths":303,"source":276,"sourceLanguage":18,"type":251},"skills/sound-fx","Use this skill whenever the user wants to generate sound effects, ambient audio, or short audio clips from a text description. Triggers include: any mention of 'sound effect', 'sfx', 'generate sound', 'make a sound', 'audio effect', 'ambient sound', 'foley', 'sound clip', 'noise', or requests to produce a specific sound (e.g. 'make a gunshot sound', 'generate thunder', 'create the sound of rain'). Also use when the user describes an action or scenario and wants the corresponding audio (e.g. 'someone getting spanked', 'a door slamming', 'cartoon boing'). Do NOT use for speech synthesis, music generation with melody/lyrics, or voice cloning.","sound-fx",{"claudeCode":12},"SKILL.md frontmatter at skills/sound-fx/SKILL.md",[304,305,308],{"path":271,"priority":272},{"path":306,"priority":307},"ref_3rd_party.md","medium",{"path":309,"priority":275},"scripts/sfx.py",{"basePath":311,"description":312,"displayName":313,"installMethods":314,"rationale":315,"selectedPaths":316,"source":276,"sourceLanguage":18,"type":251},"skills/speech-to-text","Use this skill whenever the user wants to transcribe audio to text, convert speech to text, or get a transcript from an audio or video file. Triggers include: any mention of 'transcribe', 'transcription', 'speech to text', 'STT', 'convert audio to text', 'what does this audio say', 'get transcript', 'subtitle generation', or requests to extract spoken words from a file. Also use when the user wants speaker identification from audio, timestamps for captions, or multilingual transcription.","speech-to-text",{"claudeCode":12},"SKILL.md frontmatter at skills/speech-to-text/SKILL.md",[317,318],{"path":271,"priority":272},{"path":319,"priority":275},"scripts/stt.py",{"basePath":321,"description":322,"displayName":323,"installMethods":324,"rationale":325,"selectedPaths":326,"source":276,"sourceLanguage":18,"type":251},"skills/template-skill","Reusable template for authoring new Agent Skills with clear triggers, workflow, and I/O contracts.","template-skill",{"claudeCode":12},"SKILL.md frontmatter at skills/template-skill/SKILL.md",[327,328],{"path":271,"priority":272},{"path":329,"priority":307},"SKILL.zh-CN.md",{"basePath":331,"description":332,"displayName":217,"installMethods":333,"rationale":334,"selectedPaths":335,"source":276,"sourceLanguage":18,"type":251},"skills/tts","Use this skill whenever the user wants to convert text into speech, generate audio from text, or produce voiceovers. Triggers include: any mention of 'TTS', 'text to speech', 'speak', 'say', 'voice', 'read aloud', 'audio narration', 'voiceover', 'dubbing', or requests to turn written content into spoken audio. Also use when converting EPUB/PDF/SRT/articles to audio, cloning voices from reference audio, controlling emotion or speed in speech, aligning speech to subtitle timelines, or producing per-segment voice-mapped audio.",{"claudeCode":12},"SKILL.md frontmatter at skills/tts/SKILL.md",[336,337,338,340,342,344,346],{"path":271,"priority":272},{"path":306,"priority":307},{"path":339,"priority":275},"scripts/noiz_tts.py",{"path":341,"priority":275},"scripts/render_timeline.py",{"path":343,"priority":275},"scripts/test_tts.py",{"path":345,"priority":275},"scripts/text_to_srt.py",{"path":347,"priority":275},"scripts/tts.py",{"basePath":349,"description":350,"displayName":351,"installMethods":352,"rationale":353,"selectedPaths":354,"source":276,"sourceLanguage":18,"type":251},"skills/video-translation","Translate and dub videos from one language to another, replacing the original audio with TTS while keeping the video intact.","video-translation",{"claudeCode":12},"SKILL.md frontmatter at skills/video-translation/SKILL.md",[355,356,358],{"path":271,"priority":272},{"path":357,"priority":275},"scripts/replace_audio.sh",{"path":359,"priority":275},"scripts/srt_to_duck.py",{"sources":361},[362],"manual",{"closedIssues90d":8,"description":364,"forks":238,"openIssues90d":239,"pushedAt":240,"readmeSize":236,"stars":241,"topics":365},"Allow your 🦞 bot to Shout, Speak, with \"human\" vibe",[],{"classifiedAt":367,"discoverAt":368,"extractAt":369,"githubAt":369,"updatedAt":367},1778694659166,1778694655410,1778694657421,[221,220,222,219,217,218],{"evaluatedAt":246,"extractAt":372,"updatedAt":373},1778694659798,1778694913797,[],[376,396,425,450,477,508],{"_creationTime":377,"_id":378,"community":379,"display":380,"identity":382,"providers":383,"relations":391,"tags":392,"workflow":393},1778694659798.255,"k1780m9bym6pcnr3yx322hj5ah86n6y7",{"reviewCount":8},{"description":332,"installMethods":381,"name":217,"sourceUrl":14},{"claudeCode":12},{"basePath":331,"githubOwner":249,"githubRepo":250,"locale":18,"slug":217,"type":251},{"evaluate":384,"extract":390},{"promptVersionExtension":210,"promptVersionScoring":211,"score":385,"tags":386,"targetMarket":223,"tier":224},96,[217,219,387,218,388,389],"audio","narration","dubbing",{"commitSha":257},{"repoId":259},[387,389,388,219,217,218],{"evaluatedAt":394,"extractAt":372,"updatedAt":395},1778694818344,1778694914881,{"_creationTime":397,"_id":398,"community":399,"display":400,"identity":406,"providers":410,"relations":418,"tags":421,"workflow":422},1778696833339.6235,"k173202da6bz9n3t2ytqn33f3h86n9e0",{"reviewCount":8},{"description":401,"installMethods":402,"name":404,"sourceUrl":405},"Convert documents and text to audio using Google Cloud Text-to-Speech.\nUse this skill when the user wants to: narrate a document, read aloud text,\ngenerate audio from a file, convert text to speech, create a recording\nof documentation or analysis, create a podcast from a document, or use\nGoogle TTS/text-to-speech. Trigger phrases: \"read this aloud\", \"narrate this\",\n\"create a recording\", \"text to speech\", \"TTS\", \"convert to audio\",\n\"audio from document\", \"listen to this\", \"generate audio\", \"google tts\",\n\"create a podcast\".\n",{"claudeCode":403},"sanjay3290/ai-skills","google-tts","https://github.com/sanjay3290/ai-skills",{"basePath":407,"githubOwner":408,"githubRepo":409,"locale":18,"slug":404,"type":251},"skills/google-tts","sanjay3290","ai-skills",{"evaluate":411,"extract":417},{"promptVersionExtension":210,"promptVersionScoring":211,"score":412,"tags":413,"targetMarket":223,"tier":416},100,[217,219,387,414,415],"google-cloud","python","verified",{"commitSha":257},{"parentExtensionId":419,"repoId":420},"k17es37z10n1sw6t2m3f0vsydx86mnje","kd71np0fyqg23qg8w2hcfw0h0h86nkn0",[387,414,415,219,217],{"evaluatedAt":423,"extractAt":424,"updatedAt":423},1778697061989,1778696833339,{"_creationTime":426,"_id":427,"community":428,"display":429,"identity":435,"providers":438,"relations":444,"tags":446,"workflow":447},1778695021936.5625,"k1703jnec5x3fdsmasfekwt56s86m8bv",{"reviewCount":8},{"description":430,"installMethods":431,"name":433,"sourceUrl":434},"Use when the user asks for text-to-speech narration or voiceover, accessibility reads, audio prompts, or batch speech generation via the OpenAI Audio API; run the bundled CLI (`scripts/text_to_speech.py`) with built-in voices and require `OPENAI_API_KEY` for live calls. Custom voice creation is out of scope.",{"claudeCode":432},"openai/skills","Speech Generation Skill","https://github.com/openai/skills",{"basePath":436,"githubOwner":437,"githubRepo":250,"locale":18,"slug":219,"type":251},"skills/.curated/speech","openai",{"evaluate":439,"extract":442},{"promptVersionExtension":210,"promptVersionScoring":211,"score":412,"tags":440,"targetMarket":223,"tier":416},[219,387,437,217,441,388],"text-to-speech",{"commitSha":257,"license":443},"Apache-2.0",{"repoId":445},"kd75n2zj3yh472p25zffgycved86mnpx",[387,388,437,219,441,217],{"evaluatedAt":448,"extractAt":449,"updatedAt":448},1778695652845,1778695021936,{"_creationTime":451,"_id":452,"community":453,"display":454,"identity":460,"providers":464,"relations":471,"tags":473,"workflow":474},1778686012417.834,"k17e69ew61t1yn58nb6f6d9r9586mrga",{"reviewCount":8},{"description":455,"installMethods":456,"name":458,"sourceUrl":459},"Convert speech to text using `sag` (ElevenLabs STT) and synthesize speech using `say` (macOS built-in TTS). Enables voice input transcription and audio output.",{"claudeCode":457},"daxaur/openpaw","c-voice","https://github.com/daxaur/openpaw",{"basePath":461,"githubOwner":462,"githubRepo":463,"locale":18,"slug":458,"type":251},"skills/c-voice","daxaur","openpaw",{"evaluate":465,"extract":470},{"promptVersionExtension":210,"promptVersionScoring":211,"score":466,"tags":467,"targetMarket":223,"tier":224},75,[218,219,217,468,387,469],"stt","macos",{"commitSha":257},{"repoId":472},"kd7bwdbyne8rrdfxt9ahxb6ved86mk9b",[387,469,219,468,217,218],{"evaluatedAt":475,"extractAt":476,"updatedAt":475},1778686962243,1778686012417,{"_creationTime":478,"_id":479,"community":480,"display":481,"identity":487,"providers":492,"relations":501,"tags":504,"workflow":505},1778697045057.978,"k1709qff277g3qreq668nrfj0d86nrrb",{"reviewCount":8},{"description":482,"installMethods":483,"name":485,"sourceUrl":486},"Scrub AI tells from any text draft OR audit a finished post against the 2026 algorithm heuristic checklist. Tier-based rewriter (forensic / strict / aesthetic / all) plus `--mode audit` for detection-only pass-fail review covering length, hook, CTA, format penalties, AI vocab. Sub-tools: emoji-pattern detector, multi-detector spread tester (GPTZero, Originality.ai, ZeroGPT, Sapling, Copyleaks), rule explainer. Triggers on \"humanize\", \"de-AI\", \"review this draft\", \"audit before posting\", \"is this ready\".",{"claudeCode":484},"sergebulaev/linkedin-skills","LinkedIn Humanizer","https://github.com/sergebulaev/linkedin-skills",{"basePath":488,"githubOwner":489,"githubRepo":490,"locale":18,"slug":491,"type":251},"skills/linkedin-humanizer","sergebulaev","linkedin-skills","linkedin-humanizer",{"evaluate":493,"extract":499},{"promptVersionExtension":210,"promptVersionScoring":211,"score":412,"tags":494,"targetMarket":223,"tier":416},[495,496,497,498,222],"ai-detection","text-rewriting","linkedin","content-generation",{"commitSha":257,"license":500},"MIT",{"parentExtensionId":502,"repoId":503},"k17bg3ahwtt998cj512cdc040x86m16m","kd7eh8yxvx7bp76mv9rcfdjk0x86m12a",[495,498,497,222,496],{"evaluatedAt":506,"extractAt":507,"updatedAt":506},1778697181923,1778697045058,{"_creationTime":509,"_id":510,"community":511,"display":512,"identity":518,"providers":522,"relations":531,"tags":534,"workflow":535},1778691104675.98,"k17a012kzjtmn6vm9xf7k1q3d986n6me",{"reviewCount":8},{"description":513,"installMethods":514,"name":516,"sourceUrl":517},"Convert a resume PDF to clean markdown for LLM parsing or candidate pipelines.",{"claudeCode":515},"iterationlayer/skills","Convert Resume to Markdown","https://github.com/iterationlayer/skills",{"basePath":519,"githubOwner":520,"githubRepo":250,"locale":18,"slug":521,"type":251},"skills/convert-resume-to-markdown","iterationlayer","convert-resume-to-markdown",{"evaluate":523,"extract":530},{"promptVersionExtension":210,"promptVersionScoring":211,"score":412,"tags":524,"targetMarket":223,"tier":416},[525,526,527,528,529,222],"document-processing","pdf","markdown","resume","hiring",{"commitSha":257,"license":500},{"parentExtensionId":532,"repoId":533},"k1721s0xmp59902ybtpakrrffn86n10s","kd76p4g2qmtrkgx99cnab3683d86n4g8",[525,529,527,222,526,528],{"evaluatedAt":536,"extractAt":537,"updatedAt":536},1778691474825,1778691104676]