[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"extension-skill-bytedance-podcast-generation-ko":3,"guides-for-bytedance-podcast-generation":240,"similar-k17a8rwjmp1sgctvjpwerece6d867cte":241},{"_creationTime":4,"_id":5,"children":6,"community":7,"display":9,"evaluation":20,"identity":187,"isFallback":192,"parentExtension":193,"providers":194,"relations":199,"repo":201,"workflow":237},1778053100136.245,"k17a8rwjmp1sgctvjpwerece6d867cte",[],{"reviewCount":8},0,{"description":10,"installMethods":11,"name":12,"sourceUrl":13,"tags":14},"Use this skill when the user requests to generate, create, or produce podcasts from text content. Converts written content into a two-host conversational podcast audio format with natural dialogue.",{},"Podcast Generation Skill","https://github.com/bytedance/deer-flow/tree/HEAD/skills/public/podcast-generation",[15,16,17,18,19],"podcast","audio-generation","tts","python","conversational-ai",{"_creationTime":21,"_id":22,"extensionId":5,"locale":23,"result":24,"trustSignals":176,"workflow":185},1778053169012.8296,"kn76jh7g5zagvhsa6gjykyst85867xe0","en",{"checks":25,"evaluatedAt":166,"extensionSummary":167,"promptVersionExtension":168,"promptVersionScoring":169,"rationale":170,"score":171,"summary":172,"tags":173,"targetMarket":174,"tier":175},[26,31,34,37,41,44,48,53,56,59,63,68,71,75,78,81,84,87,90,93,96,100,104,108,112,115,118,121,125,128,131,134,137,140,144,147,150,153,156,159,163],{"category":27,"check":28,"severity":29,"summary":30},"Practical Utility","Problem relevance","pass","The description clearly states the user problem: generating podcasts from text content, specifically converting written content into a two-host conversational audio format.",{"category":27,"check":32,"severity":29,"summary":33},"Unique selling proposition","The skill offers significant value over a simple prompt by providing a structured workflow for script generation, TTS synthesis, and audio mixing, including support for two hosts and multiple languages.",{"category":27,"check":35,"severity":29,"summary":36},"Production readiness","The extension is production-ready, covering the full lifecycle from script generation to final MP3 output and transcript, with clear execution instructions and handling of required environment variables.",{"category":38,"check":39,"severity":29,"summary":40},"Scope","Single responsibility principle","The skill has a single, well-defined responsibility: generating podcasts from text content, and does not extend into unrelated domains.",{"category":38,"check":42,"severity":29,"summary":43},"Description quality","The displayed description is concise, accurate, and accurately reflects the functionality described in the SKILL.md file and code.",{"category":45,"check":46,"severity":29,"summary":47},"Invocation","Scoped tools","The primary tool is a single Python script with well-defined arguments, operating on specific input and output files, which is a clear and scoped approach.",{"category":49,"check":50,"severity":51,"summary":52},"Documentation","Configuration & parameter reference","warning","The required environment variables (VOLCENGINE_TTS_APPID, VOLCENGINE_TTS_ACCESS_TOKEN) are documented, but VOLCENGINE_TTS_CLUSTER is optional and has a default, which is good. However, the script's internal parameters like `--script-file` and `--output-file` are documented, but there's no mention of their expected types or validation beyond being absolute paths.",{"category":38,"check":54,"severity":29,"summary":55},"Tool naming","The core functionality is exposed through a Python script with descriptive arguments, adhering to clear naming conventions.",{"category":38,"check":57,"severity":29,"summary":58},"Minimal I/O surface","The script takes explicit file paths as input and produces specified output files, adhering to a minimal and documented I/O surface.",{"category":60,"check":61,"severity":29,"summary":62},"License","License usability","The project includes an MIT license file, which is a permissive open-source license.",{"category":64,"check":65,"severity":66,"summary":67},"Maintenance","Commit recency","not_applicable","No commit history is available for the specific skill files provided; therefore, this check cannot be performed.",{"category":64,"check":69,"severity":51,"summary":70},"Dependency Management","The Python script uses several external libraries (argparse, base64, json, logging, os, uuid, concurrent.futures, requests, typing). While these are standard, there's no explicit dependency management file (like requirements.txt or pyproject.toml) for these within the skill's directory, and no mechanism for updating them is apparent.",{"category":72,"check":73,"severity":51,"summary":74},"Security","Secret Management","The script explicitly reads sensitive Volcengine TTS credentials from environment variables (VOLCENGINE_TTS_APPID, VOLCENGINE_TTS_ACCESS_TOKEN). While this is better than hardcoding, it doesn't prevent these secrets from being potentially exposed if the environment they are stored in is compromised or if logging unintentionally reveals them.",{"category":72,"check":76,"severity":29,"summary":77},"Injection","The script executes a Python script with arguments derived from command-line parsing. The script does not appear to load or execute untrusted external data as instructions.",{"category":72,"check":79,"severity":29,"summary":80},"Transitive Supply-Chain Grenades","The skill does not fetch remote content or execute external commands at runtime beyond the Python script itself. All dependencies are standard Python libraries.",{"category":72,"check":82,"severity":29,"summary":83},"Sandbox Isolation","The script operates within its own execution context and writes files to specified output directories within the user's data path, not directly manipulating arbitrary system paths.",{"category":72,"check":85,"severity":29,"summary":86},"Sandbox escape primitives","The Python script does not appear to contain any primitives that would allow it to escape its sandbox environment.",{"category":72,"check":88,"severity":29,"summary":89},"Data Exfiltration","The script only reads necessary environment variables for API access and writes output files to designated locations. There are no indications of unauthorized data exfiltration.",{"category":72,"check":91,"severity":29,"summary":92},"Hidden Text Tricks","The bundled files (SKILL.md and Python script) do not contain any hidden text tricks, invisible Unicode characters, or other obfuscation methods.",{"category":72,"check":94,"severity":29,"summary":95},"Opaque code execution","The Python script is provided in plain, readable source code, with no obfuscation, base64 payloads, or runtime code fetching.",{"category":97,"check":98,"severity":29,"summary":99},"Portability","Structural Assumption","The script assumes standard file paths for inputs and outputs, but these are explicitly defined and configurable via arguments, not hardcoded user-specific paths.",{"category":101,"check":102,"severity":66,"summary":103},"Trust","Issues Attention","No issue tracking information (opened/closed issues) is available for this specific skill.",{"category":105,"check":106,"severity":51,"summary":107},"Versioning","Release Management","There is no explicit versioning information (e.g., a version field in a manifest or a CHANGELOG) for this skill. The README indicates the project is at version 2.0, but this skill lacks specific versioning.",{"category":109,"check":110,"severity":51,"summary":111},"Code Execution","Validation","The Python script uses argparse for command-line arguments, which provides basic validation for argument presence and type (implicitly string for paths). However, there's no explicit validation for the *content* of the paths (e.g., ensuring they are absolute, writable, or exist) or the structure of the input JSON script beyond basic dictionary access.",{"category":72,"check":113,"severity":29,"summary":114},"Unguarded Destructive Operations","The script's primary action is generating files, which is not considered a destructive operation in this context and does not require explicit guards.",{"category":109,"check":116,"severity":29,"summary":117},"Error Handling","The Python script includes comprehensive try-except blocks, logging for errors, and raises specific exceptions for issues like missing environment variables or invalid script formats. It also handles TTS API errors gracefully.",{"category":109,"check":119,"severity":29,"summary":120},"Logging","The Python script implements logging for informational messages, warnings, and errors, aiding in auditing and debugging.",{"category":122,"check":123,"severity":66,"summary":124},"Compliance","GDPR","The skill generates audio content from text and does not appear to handle or process personal data directly.",{"category":122,"check":126,"severity":29,"summary":127},"Target market","The skill supports English and Chinese content and uses generic file paths, indicating a global target market with no specific regional restrictions.",{"category":97,"check":129,"severity":29,"summary":130},"Runtime stability","The Python script uses standard libraries and assumes a Python 3.12+ environment, which is portable. It explicitly checks for required environment variables and raises errors if they are missing.",{"category":45,"check":132,"severity":29,"summary":133},"Precise Purpose","The skill clearly states its purpose: generating podcasts from text content, converting it into a two-host conversational audio format. The input (text content) and output (podcast audio) are clearly defined.",{"category":45,"check":135,"severity":29,"summary":136},"Concise Frontmatter","The frontmatter in SKILL.md is concise, accurately describing the skill's core capability and function.",{"category":49,"check":138,"severity":29,"summary":139},"Concise Body","The SKILL.md file is well-structured and not excessively long, keeping the core instructions concise and delegating detailed examples and templates to separate sections.",{"category":141,"check":142,"severity":29,"summary":143},"Context","Progressive Disclosure","The SKILL.md file uses progressive disclosure by embedding a template file (tech-explainer.md) and providing a clear example of script JSON and generation command, rather than embedding all content inline.",{"category":141,"check":145,"severity":66,"summary":146},"Forked exploration","This skill performs a direct generation task and does not involve deep exploration or multi-file inspection, so 'context: fork' is not applicable.",{"category":27,"check":148,"severity":29,"summary":149},"Usage examples","The SKILL.md provides a comprehensive, ready-to-use example of user request, script JSON, and generation command, demonstrating the claimed output.",{"category":27,"check":151,"severity":51,"summary":152},"Edge cases","While the script handles some errors (missing env vars, invalid script format, TTS API errors), it doesn't explicitly document failure modes like malformed JSON in the script file or issues with the output file path (e.g., non-writable directory). The recovery steps for these are implicit (script failure).",{"category":109,"check":154,"severity":66,"summary":155},"Tool Fallback","This skill does not rely on external MCP servers or tools that would require fallback paths.",{"category":97,"check":157,"severity":51,"summary":158},"Stack assumptions","The Python script assumes a Python 3.12+ environment and requires specific Volcengine TTS environment variables. While Python is common, the specific TTS environment variables are a necessary precondition not explicitly declared upfront in the SKILL.md's main instruction section, though they are listed under 'Requirements'.",{"category":160,"check":161,"severity":29,"summary":162},"Safety","Halt on unexpected state","The script checks for essential prerequisites like environment variables and valid script formats, halting execution and reporting errors if these conditions are not met.",{"category":97,"check":164,"severity":29,"summary":165},"Cross-skill coupling","The skill is self-contained and does not implicitly rely on other skills. It clearly defines its inputs and outputs.",1778053140127,"It converts written content into a structured JSON script, synthesizes speech using Volcengine TTS, and combines the audio into an MP3 file, also generating a markdown transcript. The skill supports both English and Chinese languages.","2.0.0","3.4.0","This skill is well-implemented, with clear documentation and a good balance of features for its purpose. The main areas for improvement are more robust input validation for file paths and better documentation of the specific Python environment version required. The lack of explicit versioning for the skill itself also detracts slightly from its maintenance status.",90,"This skill generates podcasts from text content, creating a two-host conversational audio experience.",[15,16,17,18,19],"global","verified",{"codeQuality":177,"collectedAt":178,"documentation":179,"maintenance":181,"security":182,"testCoverage":184},{},1778053127916,{"descriptionLength":180,"readmeSize":8},197,{},{"hasNpmPackage":183,"smitheryVerified":183},false,{"hasCi":183,"hasTests":183},{"updatedAt":186},1778053169012,{"githubOwner":188,"githubRepo":189,"locale":23,"slug":190,"type":191},"bytedance","deer-flow","podcast-generation","skill",true,null,{"extract":195,"llm":198},{"commitSha":196,"license":197},"1336872b15c25d45ebcb7c1cf72369c2bdd53187","MIT",{"promptVersionExtension":168,"promptVersionScoring":169,"score":171,"targetMarket":174,"tier":175},{"repoId":200},"kd789sm7egx1h0t1jag6zzhcq98656wv",{"_creationTime":202,"_id":200,"identity":203,"providers":205,"workflow":234},1777995558409.9045,{"githubOwner":188,"githubRepo":189,"sourceUrl":204},"https://github.com/bytedance/deer-flow",{"discover":206,"github":209},{"sources":207},[208],"skills-sh",{"closedIssues90d":210,"forks":211,"homepage":212,"license":197,"openIssues90d":213,"pushedAt":214,"readmeSize":215,"stars":216,"topics":217},389,8629,"https://deerflow.tech",356,1778052455000,38642,65247,[218,219,220,221,222,223,224,225,226,227,228,229,15,18,230,231,232,233],"agent","agentic","agentic-framework","agentic-workflow","ai","ai-agents","deep-research","langchain","langgraph","llm","multi-agent","nodejs","langmanus","typescript","harness","superagent",{"discoverAt":235,"extractAt":236,"githubAt":236,"updatedAt":236},1777995558409,1778053102364,{"anyEnrichmentAt":238,"extractAt":239,"githubAt":238,"llmAt":186,"updatedAt":186},1778053101076,1778053100136,[],[242,269,298,320,356,378],{"_creationTime":243,"_id":244,"community":245,"display":246,"identity":255,"providers":258,"relations":263,"workflow":265},1778054291668.0989,"k1746xzw25syqd1mcswezg14k1867vsj",{"reviewCount":8},{"description":247,"installMethods":248,"name":249,"sourceUrl":250,"tags":251},"Use this skill whenever the user wants to convert text into speech, generate audio from text, or produce voiceovers. Triggers include: any mention of 'TTS', 'text to speech', 'speak', 'say', 'voice', 'read aloud', 'audio narration', 'voiceover', 'dubbing', or requests to turn written content into spoken audio. Also use when converting EPUB/PDF/SRT/articles to audio, cloning voices from reference audio, controlling emotion or speed in speech, aligning speech to subtitle timelines, or producing per-segment voice-mapped audio.",{},"TTS","https://github.com/noizai/skills/tree/HEAD/skills/tts",[17,252,16,253,254],"speech-synthesis","voice-cloning","dubbing",{"githubOwner":256,"githubRepo":257,"locale":23,"slug":17,"type":191},"noizai","skills",{"extract":259,"llm":261},{"commitSha":260,"license":66},"3428725f8c34e2179e0bdaf4c441697e0014d289",{"promptVersionExtension":168,"promptVersionScoring":169,"score":262,"targetMarket":174,"tier":175},95,{"repoId":264},"kd76a0e9dvgjrpvm0qe8k24qy18653ah",{"anyEnrichmentAt":266,"extractAt":267,"githubAt":266,"llmAt":268,"updatedAt":268},1778054291966,1778054291668,1778054322837,{"_creationTime":270,"_id":271,"community":272,"display":273,"identity":283,"providers":287,"relations":292,"workflow":294},1778053622473.6453,"k174xmaemtbdy32kf1f0x0rxs1867k5k",{"reviewCount":8},{"description":274,"installMethods":275,"name":276,"sourceUrl":277,"tags":278},"Convert written documents to narrated video scripts with TTS audio and word-level timing. Use when preparing essays, blog posts, or articles for video narration. Outputs scene files, audio, and VTT with precise word timestamps. Keywords: narration, voiceover, TTS, scenes, audio, timing, video script, spoken.",{},"Document to Narration","https://github.com/jwynia/agent-skills/tree/HEAD/skills/education/document-to-narration",[17,279,280,281,16,282],"narration","video-production","scripting","transcription",{"githubOwner":284,"githubRepo":285,"locale":23,"slug":286,"type":191},"jwynia","agent-skills","document-to-narration",{"extract":288,"llm":290},{"commitSha":289,"license":197},"e02ec7e226a6e4f8419fd3b88a1d8e472d421b32",{"promptVersionExtension":168,"promptVersionScoring":169,"score":291,"targetMarket":174,"tier":175},88,{"repoId":293},"kd7efn3mprpa8rd8vm5hw5ebzx864fph",{"anyEnrichmentAt":295,"extractAt":296,"githubAt":295,"llmAt":297,"updatedAt":297},1778053625386,1778053622473,1778054012696,{"_creationTime":299,"_id":300,"community":301,"display":302,"identity":310,"providers":312,"relations":318,"workflow":319},1778054291668.0974,"k17e63tha4tskpe6xrytq3qxtn867m2j",{"reviewCount":8},{"description":303,"installMethods":304,"name":305,"sourceUrl":306,"tags":307},"Fetches the latest news using news-aggregator-skill, formats it into a podcast script in Markdown format, and uses the tts skill to generate a podcast audio file. Use when the user asks to get the latest news and read it out as a podcast.",{},"Daily News Caster","https://github.com/noizai/skills/tree/HEAD/skills/daily-news-caster",[308,15,17,281,309],"news","automation",{"githubOwner":256,"githubRepo":257,"locale":23,"slug":311,"type":191},"daily-news-caster",{"extract":313,"llm":315},{"commitSha":260,"license":314},"n/a",{"promptVersionExtension":168,"promptVersionScoring":169,"score":316,"targetMarket":174,"tier":317},55,"flagged",{"repoId":264},{"anyEnrichmentAt":266,"extractAt":267,"githubAt":266,"llmAt":268,"updatedAt":268},{"_creationTime":321,"_id":322,"community":323,"display":324,"identity":340,"providers":344,"relations":349,"workflow":352},1778053197391.382,"k170wvt5rx3c1hv9a5sfkyezc1866k0q",{"reviewCount":8},{"description":325,"installMethods":326,"name":327,"sourceUrl":328,"tags":329},"Universal AI voice / text-to-speech skill supporting OpenAI TTS (gpt-4o-mini-tts, tts-1), ElevenLabs multilingual TTS with voice cloning, Bailian Qwen TTS (qwen-tts / qwen3-tts-vd with voice-design custom voices, long-text chunking built in), MiniMax speech-02-hd, SiliconFlow CosyVoice / SenseVoice, and PlayHT 2.0. Use this skill whenever the user asks to read text aloud, synthesize speech, generate narration, create voice-over, dub a script, or turn any text into audio (mp3 / wav / ogg / flac). Typical phrases include \"read this aloud\", \"generate voice for ...\", \"create a narration of ...\", \"tts this\", \"把这段念出来\", \"做个配音\", \"合成语音\", or mentions of voices / TTS model names like Alloy, Ash, Cherry, Rachel, CosyVoice, PlayHT. Always use this skill even if the user does not specify a provider — pick one from EXTEND.md defaults or available env keys.",{},"Happy Audio Gen","https://github.com/iamzhihuix/happy-claude-skills/tree/HEAD/skills/happy-audio-gen",[17,330,331,332,333,334,335,336,337,338,339],"speech","audio","voice-generation","openai","elevenlabs","bailian","minimax","siliconflow","playht","bun",{"githubOwner":341,"githubRepo":342,"locale":23,"slug":343,"type":191},"iamzhihuix","happy-claude-skills","happy-audio-gen",{"extract":345,"llm":347},{"commitSha":346,"license":197},"f49e7782a551759c9f9e0a4d4417ff053f0a86fd",{"promptVersionExtension":168,"promptVersionScoring":169,"score":348,"targetMarket":174,"tier":175},100,{"parentExtensionId":350,"repoId":351},"k173ydbbp6c0vdpxv5r0q9yvgd867en5","kd7dbbtdq95nkcs3k7fg9w6fdn864j0b",{"anyEnrichmentAt":353,"extractAt":354,"githubAt":353,"llmAt":355,"updatedAt":355},1778053199195,1778053197391,1778053284450,{"_creationTime":357,"_id":358,"community":359,"display":360,"identity":370,"providers":372,"relations":376,"workflow":377},1778054291668.0964,"k17477k0d5jgrwmqqkgp1phgen866eck",{"reviewCount":8},{"description":361,"installMethods":362,"name":363,"sourceUrl":364,"tags":365},"Use this skill whenever the user wants speech to sound more human, companion-like, or emotionally expressive. Triggers include: any mention of 'say like', 'talk like', 'speak like', 'companion voice', 'comfort me', 'cheer me up', 'sound more human', 'good night voice', 'good morning voice', or requests to add fillers, emotion, or personality to generated speech. Also use when the user wants to mimic a specific character's voice, apply speaking style presets (goodnight, morning, comfort, celebration, chatting), tune emotional parameters like warmth or tenderness, or make TTS output feel like a real person talking. If the user asks for a 'voice message', 'companion audio', 'character voice', or wants speech that sighs, laughs, hesitates, or sounds genuinely warm, use this skill. Do NOT use for plain text-to-speech without personality, music generation, sound effects, or general coding tasks unrelated to expressive speech.",{},"Characteristic Voice","https://github.com/noizai/skills/tree/HEAD/skills/characteristic-voice",[17,330,366,367,368,369],"voice","emotion","companion","script",{"githubOwner":256,"githubRepo":257,"locale":23,"slug":371,"type":191},"characteristic-voice",{"extract":373,"llm":374},{"commitSha":260,"license":66},{"promptVersionExtension":168,"promptVersionScoring":169,"score":375,"targetMarket":174,"tier":175},98,{"repoId":264},{"anyEnrichmentAt":266,"extractAt":267,"githubAt":266,"llmAt":268,"updatedAt":268},{"_creationTime":379,"_id":380,"community":381,"display":382,"identity":393,"providers":397,"relations":405,"workflow":407},1778053968286.4954,"k179afn14fzy4sejmjf82fgqa9867ck6",{"reviewCount":8},{"description":383,"installMethods":384,"name":385,"sourceUrl":386,"tags":387},"Searches arXiv for preprints and academic papers, retrieves abstracts, and filters by topic. Use when the user asks to find research papers, search arXiv, look up preprints, find academic articles in physics, math, CS, biology, statistics, or related fields.",{},"arXiv Search","https://github.com/langchain-ai/deepagents/tree/HEAD/libs/cli/examples/skills/arxiv-search",[388,389,390,391,18,392],"research","data-analytics","arxiv","academic-papers","cli",{"githubOwner":394,"githubRepo":395,"locale":23,"slug":396,"type":191},"langchain-ai","deepagents","arxiv-search",{"extract":398,"llm":400,"smithery":401},{"commitSha":399,"license":197},"b108c71d0c570e16c7050c1eac482e15dc35a5ed",{"promptVersionExtension":168,"promptVersionScoring":169,"score":375,"targetMarket":174,"tier":175},{"qualityScore":402,"totalActivations":403,"uniqueUsers":404,"useCount":8,"verified":183},0.7439563,9,8,{"repoId":406},"kd76dna2fvfbnjvzcpd2cwqnyd865xz7",{"anyEnrichmentAt":408,"extractAt":409,"githubAt":410,"llmAt":411,"smitheryAt":408,"updatedAt":411},1778053994907,1778053968286,1778053969344,1778054053159]