[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"extension-skill-firecrawl-firecrawl-download-en":3,"guides-for-firecrawl-firecrawl-download":506,"similar-k17bexzyshtx1ecvmz2x6728nn86n0fa-en":507},{"_creationTime":4,"_id":5,"children":6,"community":7,"display":9,"evaluation":15,"identity":259,"isFallback":249,"parentExtension":264,"providers":316,"relations":320,"repo":321,"tags":504,"workflow":505},1778686940775.5715,"k17bexzyshtx1ecvmz2x6728nn86n0fa",[],{"reviewCount":8},0,{"description":10,"installMethods":11,"name":13,"sourceUrl":14},"Download an entire website as local files — markdown, screenshots, or multiple formats per page. Use this skill when the user wants to save a site locally, download documentation for offline use, bulk-save pages as files, or says \"download the site\", \"save as local files\", \"offline copy\", \"download all the docs\", or \"save for reference\". Combines site mapping and scraping into organized local directories.\n",{"claudeCode":12},"firecrawl/cli","Firecrawl CLI","https://github.com/firecrawl/cli",{"_creationTime":16,"_id":17,"extensionId":5,"locale":18,"result":19,"trustSignals":239,"workflow":257},1778687082294.283,"kn7bgmrht23f0a3s1k583e2gq586ng3n","en",{"checks":20,"evaluatedAt":193,"extensionSummary":194,"features":195,"nonGoals":202,"practices":206,"prerequisites":210,"promptVersionExtension":213,"promptVersionScoring":214,"purpose":215,"rationale":216,"score":217,"summary":218,"tags":219,"targetMarket":225,"tier":226,"useCases":227,"workflow":232},[21,26,29,32,36,39,43,47,50,53,57,61,64,68,71,74,77,80,83,86,90,94,99,103,107,110,113,116,120,123,126,129,132,135,138,142,146,150,154,158,161,164,167,170,174,177,180,183,186,190],{"category":22,"check":23,"severity":24,"summary":25},"Practical Utility","Problem relevance","pass","The description clearly names the problem of needing to save entire websites locally and provides specific user intents and keywords.",{"category":22,"check":27,"severity":24,"summary":28},"Unique selling proposition","The skill uniquely combines site mapping and scraping into organized local directories for bulk saving, offering value beyond basic web scraping.",{"category":22,"check":30,"severity":24,"summary":31},"Production readiness","The CLI tool is fully implemented and covers the entire lifecycle of downloading a website, including options for filtering and format selection.",{"category":33,"check":34,"severity":24,"summary":35},"Scope","Single responsibility principle","The extension focuses solely on downloading website content and related operations, without extending into unrelated domains.",{"category":33,"check":37,"severity":24,"summary":38},"Description quality","The displayed description accurately reflects the functionality of downloading entire websites with various formats and options.",{"category":40,"check":41,"severity":24,"summary":42},"Invocation","Scoped tools","The CLI uses specific commands like `download`, `scrape`, `search`, `map`, `crawl`, `agent`, `credit-usage`, `config`, `view-config`, `login`, and `logout`, rather than a single generalist execution tool.",{"category":44,"check":45,"severity":24,"summary":46},"Documentation","Configuration & parameter reference","All command-line options and parameters for the various commands are clearly documented with descriptions and examples.",{"category":33,"check":48,"severity":24,"summary":49},"Tool naming","Tool names like `download`, `scrape`, `search`, `map`, and `crawl` are descriptive and relevant to their functions.",{"category":33,"check":51,"severity":24,"summary":52},"Minimal I/O surface","Command-line arguments are well-defined and specific. Output formats are clearly delineated, and options like `--output` allow targeted saving.",{"category":54,"check":55,"severity":24,"summary":56},"License","License usability","The extension is distributed under the MIT license, which is permissive and standard for open-source software.",{"category":58,"check":59,"severity":24,"summary":60},"Maintenance","Commit recency","The latest commit was on May 12, 2026, indicating recent maintenance.",{"category":58,"check":62,"severity":24,"summary":63},"Dependency Management","The `package.json` indicates dependencies are managed, and the presence of `npm install -g firecrawl-cli` suggests a standard installation process.",{"category":65,"check":66,"severity":24,"summary":67},"Security","Secret Management","The tool handles API keys via environment variables or prompts, and does not appear to log or expose secrets in output.",{"category":65,"check":69,"severity":24,"summary":70},"Injection","The tool primarily interacts with web content and APIs; there's no indication of loading or executing arbitrary third-party code as instructions.",{"category":65,"check":72,"severity":24,"summary":73},"Transitive Supply-Chain Grenades","The CLI installs via npm and primarily interacts with Firecrawl's own API, with no evidence of fetching or executing arbitrary remote code.",{"category":65,"check":75,"severity":24,"summary":76},"Sandbox Isolation","As a CLI tool, its operations are generally confined to network requests and local file writes within specified output paths, without modifying system files outside its scope.",{"category":65,"check":78,"severity":24,"summary":79},"Sandbox escape primitives","No evidence of detached process spawns or deny-retry loops in the CLI's command structures.",{"category":65,"check":81,"severity":24,"summary":82},"Data Exfiltration","The tool's primary function involves fetching public web data. Telemetry is opt-in and anonymized, and no evidence suggests exfiltration of confidential data.",{"category":65,"check":84,"severity":24,"summary":85},"Hidden Text Tricks","The documentation and code appear to be free of hidden text tricks or malicious Unicode characters.",{"category":87,"check":88,"severity":24,"summary":89},"Hooks","Opaque code execution","The CLI is installed via npm, and its commands are standard executables, not obfuscated code.",{"category":91,"check":92,"severity":24,"summary":93},"Portability","Structural Assumption","The CLI operates on URLs and saves files to specified locations, without making assumptions about the user's project directory structure beyond the output path.",{"category":95,"check":96,"severity":97,"summary":98},"Trust","Issues Attention","info","Open issues: 4, Closed issues (last 90d): 1. The closure rate is low (20%), suggesting maintainers may be slow to address issues.",{"category":100,"check":101,"severity":24,"summary":102},"Versioning","Release Management","The CLI version is indicated in `package.json` and `firecrawl --version`, and installation instructions can specify versions or use global installs.",{"category":104,"check":105,"severity":24,"summary":106},"Code Execution","Validation","Command-line arguments are parsed, and options like `--format`, `--limit`, and `--schema` imply validation and sanitization of inputs.",{"category":65,"check":108,"severity":24,"summary":109},"Unguarded Destructive Operations","While the tool modifies files, destructive operations like `rm -rf` are not exposed. File writes are generally to specified output directories and can be controlled by the user.",{"category":104,"check":111,"severity":24,"summary":112},"Error Handling","The CLI provides clear error messages for invalid commands, authentication issues, and API errors, often with suggested next steps.",{"category":104,"check":114,"severity":24,"summary":115},"Logging","The tool provides optional output to stdout and file (`-o`), allowing users to review actions. Telemetry is opt-in and anonymized.",{"category":117,"check":118,"severity":24,"summary":119},"Compliance","GDPR","The tool primarily processes public web data and does not appear to handle personal data without explicit user action or consent.",{"category":117,"check":121,"severity":24,"summary":122},"Target market","The tool's functionality is global, focusing on web scraping and data extraction without regional restrictions.",{"category":91,"check":124,"severity":24,"summary":125},"Runtime stability","As a Node.js CLI, it should be stable across POSIX-compliant systems and Windows, with clear error messages if specific runtimes are needed.",{"category":44,"check":127,"severity":24,"summary":128},"README","The README is comprehensive, detailing installation, authentication, commands, options, and examples.",{"category":33,"check":130,"severity":24,"summary":131},"Tool surface size","The CLI exposes a reasonable number of distinct commands (scrape, search, map, crawl, agent, etc.) for its domain.",{"category":40,"check":133,"severity":24,"summary":134},"Overlapping near-synonym tools","Commands like `scrape`, `map`, `crawl`, and `agent` have distinct purposes and do not appear to be near-synonyms.",{"category":44,"check":136,"severity":24,"summary":137},"Phantom features","All advertised features, such as download, search, and scraping, are implemented and documented within the CLI commands.",{"category":139,"check":140,"severity":24,"summary":141},"Install","Installation instruction","Clear installation instructions are provided via npm, with examples for basic usage, authentication, and setup.",{"category":143,"check":144,"severity":24,"summary":145},"Errors","Actionable error messages","Error messages are generally clear and provide guidance on how to resolve issues, such as authentication or invalid parameters.",{"category":147,"check":148,"severity":24,"summary":149},"Execution","Pinned dependencies","The `package.json` likely pins dependencies, and `npm install -g` ensures a specific version is installed.",{"category":33,"check":151,"severity":152,"summary":153},"Dry-run preview","not_applicable","While file writes occur, there isn't a direct 'dry-run' flag for all operations. However, commands like `map` preview URLs before scraping, and the overall design focuses on data retrieval and organization rather than destructive actions.",{"category":155,"check":156,"severity":24,"summary":157},"Protocol","Idempotent retry & timeouts","The tool interacts with web services, which may have timeouts. Options like `--timeout` are available for some operations, and the nature of scraping/downloading generally allows for retries.",{"category":117,"check":159,"severity":24,"summary":160},"Telemetry opt-in","Telemetry is explicitly opt-in via `FIRECRAWL_NO_TELEMETRY=1` and is anonymized, collecting only basic usage and environment data.",{"category":40,"check":162,"severity":24,"summary":163},"Precise Purpose","The description clearly states the purpose (downloading websites locally) and provides specific use cases and trigger phrases.",{"category":40,"check":165,"severity":24,"summary":166},"Concise Frontmatter","The frontmatter is concise and effectively summarizes the core capability of downloading websites, with trigger phrases.",{"category":44,"check":168,"severity":24,"summary":169},"Concise Body","The SKILL.md is well-structured, with commands documented separately and details like examples and options provided concisely.",{"category":171,"check":172,"severity":24,"summary":173},"Context","Progressive Disclosure","Detailed information, such as command options and examples, are presented clearly within the documentation without overwhelming the core description.",{"category":171,"check":175,"severity":152,"summary":176},"Forked exploration","This is a CLI tool and does not involve deep exploration requiring a forked context.",{"category":22,"check":178,"severity":24,"summary":179},"Usage examples","The documentation provides numerous ready-to-use examples for each command, covering various options and use cases.",{"category":22,"check":181,"severity":24,"summary":182},"Edge cases","The documentation lists limitations and options for handling various scenarios, like rate limiting, timeouts, and specific file formats.",{"category":104,"check":184,"severity":152,"summary":185},"Tool Fallback","The skill does not rely on external MCP servers or tools that would require a fallback.",{"category":187,"check":188,"severity":24,"summary":189},"Safety","Halt on unexpected state","The CLI provides clear error messages for invalid states or parameters, halting execution and guiding the user.",{"category":91,"check":191,"severity":24,"summary":192},"Cross-skill coupling","The tool operates as a standalone CLI and does not implicitly rely on other skills.",1778687082175,"This is a command-line interface tool that allows users to download entire websites as local files in various formats (markdown, screenshots, etc.), search the web, scrape individual pages, crawl websites, and run AI agent tasks for data extraction.",[196,197,198,199,200,201],"Download entire websites as local files","Scrape individual web pages in multiple formats","Search the web with various filters","Crawl websites for content","AI-powered web data extraction agent","Command-line interface for web interaction",[203,204,205],"Performing real-time web application interaction beyond scraping","Replacing full-fledged browser automation frameworks for complex UI testing","Storing sensitive user credentials insecurely",[207,208,209],"Web Scraping","Data Extraction","Command-line Utility",[211,212],"Node.js installed","npm package manager","3.0.0","4.4.0","To enable users to efficiently save website content locally for offline access, research, or bulk data extraction.","The extension is production-ready, well-documented, and covers its scope comprehensively. Minor info finding on issue closure rate does not significantly detract from overall quality.",98,"A robust and well-documented CLI tool for downloading websites and extracting web content.",[220,221,222,223,224],"web-scraping","download","cli","data-extraction","automation","global","verified",[228,229,230,231],"Saving documentation for offline use","Bulk-saving web pages for reference","Extracting structured data from websites","Automating web research tasks",[233,234,235,236,237,238],"User invokes a firecrawl command (e.g., `download`, `scrape`, `search`).","User provides necessary arguments like URL, format, or search query.","The CLI processes the request, potentially authenticating with the Firecrawl API.","The tool fetches data from the web or interacts with the Firecrawl API.","Results are displayed on stdout, saved to a file, or processed by the agent.","User reviews the output or continues with further commands.",{"codeQuality":240,"collectedAt":242,"documentation":243,"maintenance":246,"popularity":253,"security":255,"testCoverage":256},{"hasLockfile":241},true,1778687057140,{"descriptionLength":244,"readmeSize":245},409,25712,{"closedIssues90d":247,"forks":248,"hasChangelog":249,"openIssues90d":250,"pushedAt":251,"stars":252},1,49,false,4,1778599393000,383,{"npmDownloads":254},51144,{"hasNpmPackage":241,"smitheryVerified":249},{"hasCi":241,"hasTests":241},{"updatedAt":258},1778687082294,{"basePath":260,"githubOwner":261,"githubRepo":222,"locale":18,"slug":262,"type":263},"skills/firecrawl-download","firecrawl","firecrawl-download","skill",{"_creationTime":265,"_id":266,"community":267,"display":268,"identity":271,"parentExtension":274,"providers":303,"relations":312,"tags":313,"workflow":314},1778686940775.5706,"k17axfavjpz72cd3qqzn86shb186ncqt",{"reviewCount":8},{"description":269,"installMethods":270,"name":261,"sourceUrl":14},"Scrape, search, crawl, and map the web with a single command.",{"claudeCode":261},{"basePath":272,"githubOwner":261,"githubRepo":222,"locale":18,"slug":222,"type":273},"","plugin",{"_creationTime":275,"_id":276,"community":277,"display":278,"identity":281,"providers":283,"relations":297,"tags":299,"workflow":300},1778686940775.5703,"k17be4khad1zr773wdg2m4yjhn86mjpr",{"reviewCount":8},{"description":279,"installMethods":280,"name":13,"sourceUrl":14},"CLI and Agent Skill for Firecrawl - Add scrape, search, and browsing capabilities to your AI agents",{"claudeCode":12},{"basePath":272,"githubOwner":261,"githubRepo":222,"locale":18,"slug":222,"type":282},"marketplace",{"evaluate":284,"extract":291},{"promptVersionExtension":285,"promptVersionScoring":214,"score":286,"tags":287,"targetMarket":225,"tier":226},"3.1.0",97,[220,222,288,223,289,290,224],"ai-agents","search","crawl",{"commitSha":292,"license":293,"marketplace":294,"plugin":295},"HEAD","MIT",{"name":261,"pluginCount":247},{"mcpCount":8,"provider":296,"skillCount":8},"classify",{"repoId":298},"kd7csd1wb06dg9c1jfy5063f2586ne60",[288,224,222,290,223,289,220],{"evaluatedAt":301,"extractAt":302,"updatedAt":301},1778686962954,1778686940775,{"evaluate":304,"extract":309},{"promptVersionExtension":213,"promptVersionScoring":214,"score":305,"tags":306,"targetMarket":225,"tier":308},78,[220,222,224,223,289,290,307],"agent","community",{"commitSha":292,"plugin":310},{"mcpCount":8,"provider":296,"skillCount":311},9,{"parentExtensionId":276,"repoId":298},[307,224,222,290,223,289,220],{"evaluatedAt":315,"extractAt":302,"updatedAt":315},1778686985594,{"evaluate":317,"extract":319},{"promptVersionExtension":213,"promptVersionScoring":214,"score":217,"tags":318,"targetMarket":225,"tier":226},[220,221,222,223,224],{"commitSha":292,"license":293},{"parentExtensionId":266,"repoId":298},{"_creationTime":322,"_id":298,"identity":323,"providers":324,"workflow":499},1778686934511.6108,{"githubOwner":261,"githubRepo":222,"sourceUrl":14},{"classify":325,"discover":491,"extract":494,"github":495,"npm":498},{"commitSha":292,"extensions":326},[327,337,363,372,379,387,392,400,408,416,424,432],{"basePath":272,"displayName":261,"installMethods":328,"rationale":329,"selectedPaths":330,"source":336,"sourceLanguage":18,"type":282},{"claudeCode":12},"marketplace.json at .claude-plugin/marketplace.json",[331,334],{"path":332,"priority":333},".claude-plugin/marketplace.json","mandatory",{"path":335,"priority":333},"README.md","rule",{"basePath":272,"description":269,"displayName":261,"installMethods":338,"rationale":339,"selectedPaths":340,"source":336,"sourceLanguage":18,"type":273},{"claudeCode":261},"plugin manifest at .claude-plugin/plugin.json",[341,343,344,347,349,351,353,355,357,359,361],{"path":342,"priority":333},".claude-plugin/plugin.json",{"path":335,"priority":333},{"path":345,"priority":346},"skills/firecrawl-agent/SKILL.md","medium",{"path":348,"priority":346},"skills/firecrawl-cli/SKILL.md",{"path":350,"priority":346},"skills/firecrawl-crawl/SKILL.md",{"path":352,"priority":346},"skills/firecrawl-download/SKILL.md",{"path":354,"priority":346},"skills/firecrawl-interact/SKILL.md",{"path":356,"priority":346},"skills/firecrawl-map/SKILL.md",{"path":358,"priority":346},"skills/firecrawl-parse/SKILL.md",{"path":360,"priority":346},"skills/firecrawl-scrape/SKILL.md",{"path":362,"priority":346},"skills/firecrawl-search/SKILL.md",{"basePath":364,"description":365,"displayName":366,"installMethods":367,"rationale":368,"selectedPaths":369,"source":336,"sourceLanguage":18,"type":263},"skills/firecrawl-agent","AI-powered autonomous data extraction that navigates complex sites and returns structured JSON. Use this skill when the user wants structured data from websites, needs to extract pricing tiers, product listings, directory entries, or any data as JSON with a schema. Triggers on \"extract structured data\", \"get all the products\", \"pull pricing info\", \"extract as JSON\", or when the user provides a JSON schema for website data. More powerful than simple scraping for multi-page structured extraction.\n","firecrawl-agent",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-agent/SKILL.md",[370],{"path":371,"priority":333},"SKILL.md",{"basePath":373,"description":374,"displayName":261,"installMethods":375,"rationale":376,"selectedPaths":377,"source":336,"sourceLanguage":18,"type":263},"skills/firecrawl-cli","Search, scrape, and interact with the web via the Firecrawl CLI. Use this skill whenever the user wants to search the web, find articles, research a topic, look something up online, scrape a webpage, grab content from a URL, get data from a website, crawl documentation, download a site, or interact with pages that need clicks or logins. Also use when they say \"fetch this page\", \"pull the content from\", \"get the page at https://\", or reference external websites. This provides real-time web search with full page content and interact capabilities — beyond what Claude can do natively with built-in tools. Do NOT trigger for local file operations, git commands, deployments, or code editing tasks.\n",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-cli/SKILL.md",[378],{"path":371,"priority":333},{"basePath":380,"description":381,"displayName":382,"installMethods":383,"rationale":384,"selectedPaths":385,"source":336,"sourceLanguage":18,"type":263},"skills/firecrawl-crawl","Bulk extract content from an entire website or site section. Use this skill when the user wants to crawl a site, extract all pages from a docs section, bulk-scrape multiple pages following links, or says \"crawl\", \"get all the pages\", \"extract everything under /docs\", \"bulk extract\", or needs content from many pages on the same site. Handles depth limits, path filtering, and concurrent extraction.\n","firecrawl-crawl",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-crawl/SKILL.md",[386],{"path":371,"priority":333},{"basePath":260,"description":10,"displayName":262,"installMethods":388,"rationale":389,"selectedPaths":390,"source":336,"sourceLanguage":18,"type":263},{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-download/SKILL.md",[391],{"path":371,"priority":333},{"basePath":393,"description":394,"displayName":395,"installMethods":396,"rationale":397,"selectedPaths":398,"source":336,"sourceLanguage":18,"type":263},"skills/firecrawl-interact","Control and interact with a live browser session on any scraped page — click buttons, fill forms, navigate flows, and extract data using natural language prompts or code. Use when the user needs to interact with a webpage beyond simple scraping: logging into a site, submitting forms, clicking through pagination, handling infinite scroll, navigating multi-step checkout or wizard flows, or when a regular scrape failed because content is behind JavaScript interaction. Also useful for authenticated scraping via profiles. Triggers on \"interact\", \"click\", \"fill out the form\", \"log in to\", \"sign in\", \"submit\", \"paginated\", \"next page\", \"infinite scroll\", \"interact with the page\", \"navigate to\", \"open a session\", or \"scrape failed\".\n","firecrawl-interact",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-interact/SKILL.md",[399],{"path":371,"priority":333},{"basePath":401,"description":402,"displayName":403,"installMethods":404,"rationale":405,"selectedPaths":406,"source":336,"sourceLanguage":18,"type":263},"skills/firecrawl-map","Discover and list all URLs on a website, with optional search filtering. Use this skill when the user wants to find a specific page on a large site, list all URLs, see the site structure, find where something is on a domain, or says \"map the site\", \"find the URL for\", \"what pages are on\", or \"list all pages\". Essential when the user knows which site but not which exact page.\n","firecrawl-map",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-map/SKILL.md",[407],{"path":371,"priority":333},{"basePath":409,"description":410,"displayName":411,"installMethods":412,"rationale":413,"selectedPaths":414,"source":336,"sourceLanguage":18,"type":263},"skills/firecrawl-parse","Efficiently extract and convert the contents of any local file—such as PDF, DOCX, DOC, ODT, RTF, XLSX, XLS, or HTML—into clean, well-formatted markdown saved to disk. Use this skill whenever the user requests to parse, read, or extract information from a file on their computer, including phrases like “parse this PDF”, “convert this document”, “read this file”, “extract text from”, or when a local file path (not a URL) is provided. This skill offers advanced options like generating AI-powered summaries and answering questions based on the file's content. Prefer this tool over `scrape` when handling local files to deliver precise, structured outputs for downstream tasks.\n","firecrawl-parse",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-parse/SKILL.md",[415],{"path":371,"priority":333},{"basePath":417,"description":418,"displayName":419,"installMethods":420,"rationale":421,"selectedPaths":422,"source":336,"sourceLanguage":18,"type":263},"skills/firecrawl-scrape","Extract clean markdown from any URL, including JavaScript-rendered SPAs. Use this skill whenever the user provides a URL and wants its content, says \"scrape\", \"grab\", \"fetch\", \"pull\", \"get the page\", \"extract from this URL\", or \"read this webpage\". Handles JS-rendered pages, multiple concurrent URLs, and returns LLM-optimized markdown. Use this instead of WebFetch for any webpage content extraction.\n","firecrawl-scrape",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-scrape/SKILL.md",[423],{"path":371,"priority":333},{"basePath":425,"description":426,"displayName":427,"installMethods":428,"rationale":429,"selectedPaths":430,"source":336,"sourceLanguage":18,"type":263},"skills/firecrawl-search","Web search with full page content extraction. Use this skill whenever the user asks to search the web, find articles, research a topic, look something up, find recent news, discover sources, or says \"search for\", \"find me\", \"look up\", \"what are people saying about\", or \"find articles about\". Returns real search results with optional full-page markdown — not just snippets. Provides capabilities beyond Claude's built-in WebSearch.\n","firecrawl-search",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-search/SKILL.md",[431],{"path":371,"priority":333},{"basePath":272,"description":433,"displayName":434,"installMethods":435,"license":436,"rationale":437,"selectedPaths":438,"source":336,"sourceLanguage":18,"type":222},"Command-line interface for Firecrawl. Scrape, crawl, and extract data from any website directly from your terminal.","firecrawl-cli",{"npm":434},"ISC","cli ecosystem detected at /",[439,441,442,445,447,449,451,453,455,457,459,461,463,465,467,469,471,473,475,477,479,481,483,485,487,489],{"path":440,"priority":333},"package.json",{"path":335,"priority":333},{"path":443,"priority":444},"src/index.ts","low",{"path":446,"priority":444},"src/commands/agent.ts",{"path":448,"priority":444},"src/commands/browser.ts",{"path":450,"priority":444},"src/commands/config.ts",{"path":452,"priority":444},"src/commands/crawl.ts",{"path":454,"priority":444},"src/commands/create.ts",{"path":456,"priority":444},"src/commands/credit-usage.ts",{"path":458,"priority":444},"src/commands/env.ts",{"path":460,"priority":444},"src/commands/experimental/backends.ts",{"path":462,"priority":444},"src/commands/experimental/index.ts",{"path":464,"priority":444},"src/commands/experimental/shared.ts",{"path":466,"priority":444},"src/commands/init.ts",{"path":468,"priority":444},"src/commands/interact.ts",{"path":470,"priority":444},"src/commands/login.ts",{"path":472,"priority":444},"src/commands/logout.ts",{"path":474,"priority":444},"src/commands/map.ts",{"path":476,"priority":444},"src/commands/parse.ts",{"path":478,"priority":444},"src/commands/scrape.ts",{"path":480,"priority":444},"src/commands/search.ts",{"path":482,"priority":444},"src/commands/setup.ts",{"path":484,"priority":444},"src/commands/skills-install.ts",{"path":486,"priority":444},"src/commands/skills-native.ts",{"path":488,"priority":444},"src/commands/status.ts",{"path":490,"priority":444},"src/commands/version.ts",{"sources":492},[493],"manual",{"npmPackage":434},{"closedIssues90d":247,"description":279,"forks":248,"homepage":496,"openIssues90d":250,"pushedAt":251,"readmeSize":245,"stars":252,"topics":497},"http://docs.firecrawl.dev/cli",[],{"downloads":254},{"classifiedAt":500,"discoverAt":501,"extractAt":502,"githubAt":502,"npmAt":503,"updatedAt":500},1778686940560,1778686934511,1778686936677,1778686938425,[224,222,223,221,220],{"evaluatedAt":258,"extractAt":302,"updatedAt":258},[],[508,536,567,589,615,635],{"_creationTime":509,"_id":510,"community":511,"display":512,"identity":518,"providers":522,"relations":529,"tags":531,"workflow":532},1778697205743.894,"k17ej9360g7468qswf3svnrx9186m1ng",{"reviewCount":8},{"description":513,"installMethods":514,"name":516,"sourceUrl":517},"Browser automation CLI for AI agents. Use when the user needs to interact with websites, including navigating pages, filling forms, clicking buttons, taking screenshots, extracting data, testing web apps, or automating any browser task. Triggers include requests to \"open a website\", \"fill out a form\", \"click a button\", \"take a screenshot\", \"scrape data from a page\", \"test this web app\", \"login to a site\", \"automate browser actions\", or any task requiring programmatic web interaction.",{"claudeCode":515},"shanraisshan/claude-code-best-practice","agent-browser","https://github.com/shanraisshan/claude-code-best-practice",{"basePath":519,"githubOwner":520,"githubRepo":521,"locale":18,"slug":516,"type":263},".claude/skills/agent-browser","shanraisshan","claude-code-best-practice",{"evaluate":523,"extract":528},{"promptVersionExtension":213,"promptVersionScoring":214,"score":524,"tags":525,"targetMarket":225,"tier":226},100,[526,220,527,222,224],"browser-automation","testing",{"commitSha":292},{"repoId":530},"kd74710g49kxgwbfjxeb7s132d86myxr",[224,526,222,527,220],{"evaluatedAt":533,"extractAt":534,"updatedAt":535},1778697222051,1778697205743,1778697438718,{"_creationTime":537,"_id":538,"community":539,"display":540,"identity":546,"providers":551,"relations":560,"tags":563,"workflow":564},1778699234184.6157,"k177tdbfgqmwhtaqv771f2ych586nne9",{"reviewCount":8},{"description":541,"installMethods":542,"name":544,"sourceUrl":545},"Worktree-first dev environment manager for issues, PRs, and features with optional tmux sessions",{"claudeCode":543},"Yeachan-Heo/oh-my-claudecode","Project Session Manager","https://github.com/Yeachan-Heo/oh-my-claudecode",{"basePath":547,"githubOwner":548,"githubRepo":549,"locale":18,"slug":550,"type":263},"skills/project-session-manager","Yeachan-Heo","oh-my-claudecode","project-session-manager",{"evaluate":552,"extract":559},{"promptVersionExtension":213,"promptVersionScoring":214,"score":524,"tags":553,"targetMarket":225,"tier":226},[554,555,556,557,224,222,558],"git","development-environment","workflow","tmux","developer-tool",{"commitSha":292,"license":293},{"parentExtensionId":561,"repoId":562},"k17brg5egdw1jbncj1j4wfv3fh86n639","kd74zv63fryf9prygtq7gf4es986n22y",[224,222,558,555,554,557,556],{"evaluatedAt":565,"extractAt":566,"updatedAt":565},1778699613343,1778699234184,{"_creationTime":568,"_id":569,"community":570,"display":571,"identity":574,"providers":577,"relations":585,"tags":586,"workflow":587},1778699234184.611,"k179b6dkc777g1rgyecze04wqn86m6y4",{"reviewCount":8},{"description":572,"installMethods":573,"name":549,"sourceUrl":545},"Process-first advisor routing for Claude, Codex, or Gemini via `omc ask`, with artifact capture and no raw CLI assembly",{"claudeCode":543},{"basePath":575,"githubOwner":548,"githubRepo":549,"locale":18,"slug":576,"type":263},"skills/ask","ask",{"evaluate":578,"extract":584},{"promptVersionExtension":213,"promptVersionScoring":214,"score":524,"tags":579,"targetMarket":225,"tier":226},[222,224,580,581,582,583],"llm","prompting","code-review","artifact-generation",{"commitSha":292,"license":293},{"parentExtensionId":561,"repoId":562},[583,224,222,582,580,581],{"evaluatedAt":588,"extractAt":566,"updatedAt":588},1778699303045,{"_creationTime":590,"_id":591,"community":592,"display":593,"identity":599,"providers":603,"relations":609,"tags":611,"workflow":612},1778697652123.8928,"k171pew5empzzrfghyg9nqrk6n86nqa9",{"reviewCount":8},{"description":594,"installMethods":595,"name":597,"sourceUrl":598},"Use gh for GitHub issues, PR status, CI/logs, comments, reviews, releases, and API queries.",{"claudeCode":596},"steipete/clawdis","github","https://github.com/steipete/clawdis",{"basePath":600,"githubOwner":601,"githubRepo":602,"locale":18,"slug":597,"type":263},"skills/github","steipete","clawdis",{"evaluate":604,"extract":608},{"promptVersionExtension":213,"promptVersionScoring":214,"score":524,"tags":605,"targetMarket":225,"tier":226},[597,222,606,607,224],"api","developer-tools",{"commitSha":292},{"repoId":610},"kd738npxg9yh3xf3vddzy9fyfh86nhng",[606,224,222,607,597],{"evaluatedAt":613,"extractAt":614,"updatedAt":613},1778698569289,1778697652123,{"_creationTime":616,"_id":617,"community":618,"display":619,"identity":623,"providers":625,"relations":631,"tags":632,"workflow":633},1778697652123.8857,"k179k1n5nw0md1x1j3dak31evn86mb5f",{"reviewCount":8},{"description":620,"installMethods":621,"name":622,"sourceUrl":598},"Use when reviewing, reproducing, or proving OpenClaw Telegram behavior with a real Telegram user on Crabbox, including PR review workflows that need an agent-controlled Telegram Desktop recording, TDLib user-driver commands, Convex-leased credentials, WebVNC observation, and motion-trimmed artifacts.",{"claudeCode":596},"telegram-crabbox-e2e-proof",{"basePath":624,"githubOwner":601,"githubRepo":602,"locale":18,"slug":622,"type":263},".agents/skills/telegram-crabbox-e2e-proof",{"evaluate":626,"extract":630},{"promptVersionExtension":213,"promptVersionScoring":214,"score":524,"tags":627,"targetMarket":225,"tier":226},[527,628,224,629,222],"telegram","qa",{"commitSha":292},{"repoId":610},[224,222,629,628,527],{"evaluatedAt":634,"extractAt":614,"updatedAt":634},1778698038113,{"_creationTime":636,"_id":637,"community":638,"display":639,"identity":645,"providers":649,"relations":655,"tags":657,"workflow":658},1778696993586.7058,"k174ty9vbfgj7cv3v2szdwjgh986mrsf",{"reviewCount":8},{"description":640,"installMethods":641,"name":643,"sourceUrl":644},"Search ChatGPT and extract the full response + hydration JSON that powers the UI.\nAttaches to a running Chrome instance (port 9222 by default), opens ChatGPT,\nsubmits a query, waits for the streamed response, and returns structured data:\nmessages, product cards, hydration JSON, and API calls.\nUse when asked to \"search chatgpt\", \"ask chatgpt\", \"chatgpt search\",\n\"get chatgpt response\", or \"scrape chatgpt\".\n",{"claudeCode":642},"SeifBenayed/claude-code-sdk","chatgpt-search","https://github.com/SeifBenayed/claude-code-sdk",{"basePath":646,"githubOwner":647,"githubRepo":648,"locale":18,"slug":643,"type":263},".claude/skills/chatgpt-search","SeifBenayed","claude-code-sdk",{"evaluate":650,"extract":654},{"promptVersionExtension":213,"promptVersionScoring":214,"score":524,"tags":651,"targetMarket":225,"tier":226},[652,653,526,223,220],"chatgpt","scraping",{"commitSha":292},{"repoId":656},"kd78s53c1852h5p7c3qem663xs86njab",[526,652,223,653,220],{"evaluatedAt":659,"extractAt":660,"updatedAt":659},1778697026743,1778696993586]