[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"extension-skill-firecrawl-firecrawl-cli-en":3,"guides-for-firecrawl-firecrawl-cli":511,"similar-k173y2bqmv365qfxv0rebgfmxd86n50r-en":512},{"_creationTime":4,"_id":5,"children":6,"community":7,"display":9,"evaluation":15,"identity":265,"isFallback":255,"parentExtension":270,"providers":321,"relations":325,"repo":326,"tags":509,"workflow":510},1778686940775.571,"k173y2bqmv365qfxv0rebgfmxd86n50r",[],{"reviewCount":8},0,{"description":10,"installMethods":11,"name":13,"sourceUrl":14},"Search, scrape, and interact with the web via the Firecrawl CLI. Use this skill whenever the user wants to search the web, find articles, research a topic, look something up online, scrape a webpage, grab content from a URL, get data from a website, crawl documentation, download a site, or interact with pages that need clicks or logins. Also use when they say \"fetch this page\", \"pull the content from\", \"get the page at https://\", or reference external websites. This provides real-time web search with full page content and interact capabilities — beyond what Claude can do natively with built-in tools. Do NOT trigger for local file operations, git commands, deployments, or code editing tasks.\n",{"claudeCode":12},"firecrawl/cli","Firecrawl CLI","https://github.com/firecrawl/cli",{"_creationTime":16,"_id":17,"extensionId":5,"locale":18,"result":19,"trustSignals":245,"workflow":263},1778687034478.1091,"kn7cbf2d80z9j53gdh27tvmz4186mt3b","en",{"checks":20,"evaluatedAt":193,"extensionSummary":194,"features":195,"nonGoals":204,"practices":210,"prerequisites":215,"promptVersionExtension":219,"promptVersionScoring":220,"purpose":221,"rationale":222,"score":223,"summary":224,"tags":225,"targetMarket":231,"tier":232,"useCases":233,"workflow":239},[21,26,29,33,37,41,45,48,52,56,60,63,66,69,73,76,79,82,85,88,92,96,99,103,106,109,112,115,118,121,125,128,133,137,140,143,146,149,153,156,159,162,165,168,171,175,179,183,186,190],{"category":22,"check":23,"severity":24,"summary":25},"Invocation","Precise Purpose","pass","The description clearly outlines what the skill does (search, scrape, interact with the web) and when to use it, including explicit non-goals like local file operations.",{"category":22,"check":27,"severity":24,"summary":28},"Concise Frontmatter","The frontmatter is concise and effectively summarizes the core capability, followed by relevant trigger phrases.",{"category":30,"check":31,"severity":24,"summary":32},"Documentation","Concise Body","The SKILL.md file is well-structured, keeps the main instructions concise, and defers deeper material to linked files or commands.",{"category":34,"check":35,"severity":24,"summary":36},"Context","Progressive Disclosure","The skill delegates detailed procedures and command references to external files and command help, adhering to progressive disclosure.",{"category":34,"check":38,"severity":39,"summary":40},"Forked exploration","not_applicable","This skill is not primarily an exploration or audit-style skill that would benefit from forked context.",{"category":42,"check":43,"severity":24,"summary":44},"Practical Utility","Usage examples","Numerous clear and actionable examples are provided for each major capability, demonstrating input, invocation, and expected outcomes.",{"category":42,"check":46,"severity":24,"summary":47},"Edge cases","The documentation addresses various failure modes (e.g., invalid URLs, auth issues, rate limits) and limitations with recovery steps.",{"category":49,"check":50,"severity":39,"summary":51},"Code Execution","Tool Fallback","The skill does not appear to rely on external MCP servers or custom toolkits that would require a fallback.",{"category":53,"check":54,"severity":24,"summary":55},"Safety","Halt on unexpected state","The documentation and workflow suggest that operations halt on unexpected states, and output handling guidelines promote safe practices.",{"category":57,"check":58,"severity":24,"summary":59},"Portability","Cross-skill coupling","The skill is self-contained and does not implicitly rely on other skills; it cross-links to related skills explicitly when necessary.",{"category":42,"check":61,"severity":24,"summary":62},"Problem relevance","The description clearly identifies the user problem of needing real-time web search and interaction beyond native LLM capabilities.",{"category":42,"check":64,"severity":24,"summary":65},"Unique selling proposition","The Firecrawl CLI offers a significant value proposition by providing advanced web scraping, interaction, and AI-driven data extraction capabilities, going beyond basic LLM tool usage.",{"category":42,"check":67,"severity":24,"summary":68},"Production readiness","The skill is production-ready, offering a complete CLI for web interaction, with clear installation, authentication, and usage guidelines.",{"category":70,"check":71,"severity":24,"summary":72},"Scope","Single responsibility principle","The extension focuses on web interaction (search, scrape, crawl, interact, agent) and related CLI operations, maintaining a clear and coherent scope.",{"category":70,"check":74,"severity":24,"summary":75},"Description quality","The description is accurate, concise, and effectively communicates the skill's purpose and capabilities, including explicit non-goals.",{"category":22,"check":77,"severity":24,"summary":78},"Scoped tools","The skill exposes narrow, verb-noun specialist commands (e.g., `scrape`, `search`, `crawl`) rather than a single generalist tool.",{"category":30,"check":80,"severity":24,"summary":81},"Configuration & parameter reference","All command options and parameters are documented with clear descriptions in the README and SKILL.md.",{"category":70,"check":83,"severity":24,"summary":84},"Tool naming","Tool names (`scrape`, `search`, `crawl`, `map`, `agent`, `interact`) are descriptive, easy to understand, and follow kebab-case convention.",{"category":70,"check":86,"severity":24,"summary":87},"Minimal I/O surface","Tool inputs and outputs are well-defined and focused on the stated task, with options for structured output formats like JSON.",{"category":89,"check":90,"severity":24,"summary":91},"License","License usability","The project uses the MIT license, which is a permissive open-source license and is clearly stated in the repository.",{"category":93,"check":94,"severity":24,"summary":95},"Maintenance","Commit recency","The last commit was on May 12, 2026, which is within the last 3 months.",{"category":93,"check":97,"severity":24,"summary":98},"Dependency Management","The project uses npm and has a lockfile, indicating good dependency management practices.",{"category":100,"check":101,"severity":24,"summary":102},"Security","Secret Management","Secrets are handled via API keys or environment variables, with clear instructions and no evidence of secrets being echoed in output.",{"category":100,"check":104,"severity":24,"summary":105},"Injection","The CLI is designed to interact with web content and has safeguards against executing arbitrary code from fetched data.",{"category":100,"check":107,"severity":24,"summary":108},"Transitive Supply-Chain Grenades","The CLI fetches web content but does not execute it as code, mitigating risks from remote scripts or instructions.",{"category":100,"check":110,"severity":24,"summary":111},"Sandbox Isolation","The CLI operates on web content and does not appear to modify files outside of its designated output directories.",{"category":100,"check":113,"severity":24,"summary":114},"Sandbox escape primitives","No evidence of detached processes or retry loops around denied tool calls was found.",{"category":100,"check":116,"severity":24,"summary":117},"Data Exfiltration","The CLI handles web scraping and does not appear to exfiltrate confidential data; telemetry is opt-out with clear documentation.",{"category":100,"check":119,"severity":24,"summary":120},"Hidden Text Tricks","The bundled content and descriptions appear to be free of hidden steering tricks or malicious Unicode characters.",{"category":122,"check":123,"severity":24,"summary":124},"Hooks","Opaque code execution","The CLI commands are standard executable scripts and do not employ obfuscation techniques like base64 payloads or eval.",{"category":57,"check":126,"severity":24,"summary":127},"Structural Assumption","The CLI does not make structural assumptions about user project organization outside of its own output directories.",{"category":129,"check":130,"severity":131,"summary":132},"Trust","Issues Attention","warning","There were 4 issues opened and 1 issue closed in the last 90 days, indicating a closure rate below 50% and potentially slow maintainer response.",{"category":134,"check":135,"severity":24,"summary":136},"Versioning","Release Management","The CLI has a clearly defined version (e.g., v1.16.2) and is actively updated, with recent commits and npm package versions.",{"category":49,"check":138,"severity":24,"summary":139},"Validation","The CLI handles input arguments and structured output, with options for JSON schema validation for agent tasks.",{"category":100,"check":141,"severity":24,"summary":142},"Unguarded Destructive Operations","Destructive operations are not a primary function, and file writes are to designated output directories, not system-wide.",{"category":49,"check":144,"severity":24,"summary":145},"Error Handling","The CLI provides clear error messages for various scenarios, including authentication failures and invalid commands.",{"category":49,"check":147,"severity":24,"summary":148},"Logging","The CLI logs credit usage and provides status information, and telemetry is opt-in and documented.",{"category":150,"check":151,"severity":24,"summary":152},"Compliance","GDPR","The tool scrapes public web data and does not inherently operate on personal data without user-provided input or interaction.",{"category":150,"check":154,"severity":24,"summary":155},"Target market","The tool is general-purpose for web scraping and interaction and has no apparent regional or jurisdictional limitations.",{"category":57,"check":157,"severity":24,"summary":158},"Runtime stability","The CLI is a Node.js application and should be portable across POSIX-compliant systems.",{"category":30,"check":160,"severity":24,"summary":161},"README","A comprehensive README file exists and clearly states the extension's purpose, installation, and usage.",{"category":70,"check":163,"severity":24,"summary":164},"Tool surface size","The CLI exposes a reasonable number of distinct commands (scrape, search, map, crawl, agent, interact, config, login, logout, etc.).",{"category":22,"check":166,"severity":24,"summary":167},"Overlapping near-synonym tools","Tool names are distinct and cover specific functionalities, avoiding near-synonym overlap.",{"category":30,"check":169,"severity":24,"summary":170},"Phantom features","All advertised features in the README and SKILL.md have corresponding implementations in the CLI.",{"category":172,"check":173,"severity":24,"summary":174},"Install","Installation instruction","Clear installation instructions are provided via npm, along with examples for authentication and usage.",{"category":176,"check":177,"severity":24,"summary":178},"Errors","Actionable error messages","Error messages are generally clear and provide context about the failure, with instructions for authentication and setup.",{"category":180,"check":181,"severity":24,"summary":182},"Execution","Pinned dependencies","The project uses npm and likely has a lockfile, ensuring pinned dependencies.",{"category":70,"check":184,"severity":24,"summary":185},"Dry-run preview","While not explicitly a `--dry-run` flag for all operations, the output formats (like JSON) and the nature of web scraping allow for inspection of results before use.",{"category":187,"check":188,"severity":24,"summary":189},"Protocol","Idempotent retry & timeouts","Web requests have timeouts, and the CLI design focuses on stateless operations for scraping and searching.",{"category":150,"check":191,"severity":24,"summary":192},"Telemetry opt-in","Telemetry is opt-in via an environment variable and is clearly documented with limitations on collected data.",1778687034362,"The Firecrawl CLI is a command-line tool for interacting with the web, offering capabilities to search, scrape, crawl, map URLs, run AI agent jobs for data extraction, and interact with web pages. It provides clean markdown output optimized for LLM context windows and supports various formats and options for data retrieval.",[196,197,198,199,200,201,202,203],"Web search with multiple sources and filters","Scrape static and dynamic web pages","Crawl entire websites or sections","AI-powered web data extraction via agent jobs","Interactive web page manipulation","Download websites as local files","Parse local files (PDF, DOCX, etc.)","Authentication via API key or browser login",[205,206,207,208,209],"Local file operations.","Git commands.","Deployments.","Code editing tasks.","Direct manipulation of codebases.",[211,212,213,214],"Web scraping best practices","Data extraction","CLI tool usage","AI agent integration",[216,217,218],"Node.js installed","Firecrawl API key or self-hosted instance","Optional: AI coding agent (e.g., Claude Code) for skill installation","3.0.0","4.4.0","To provide users with a powerful and versatile command-line interface for accessing and extracting information from the web, going beyond the native capabilities of LLMs.","The extension is well-documented, feature-rich, and production-ready with clear installation and usage instructions. A warning regarding issue attention indicates potential maintenance slowness, capping the score.",75,"A robust and well-documented CLI for advanced web scraping, searching, and interaction.",[226,227,228,229,230],"web-scraping","cli","search","automation","data-extraction","global","community",[234,235,236,237,238],"Researching topics and finding relevant web pages.","Extracting content from specific URLs or entire websites.","Automating data collection from web sources.","Interacting with web pages that require clicks, logins, or form submissions.","Integrating web data into AI agent workflows.",[240,241,242,243,244],"Determine the user's intent (search, scrape, crawl, interact, agent).","Select the appropriate Firecrawl CLI command.","Configure command options (URL, formats, limits, schema, etc.).","Execute the command and authenticate if necessary.","Process the output (save to file, pipe to other tools, review results).",{"codeQuality":246,"collectedAt":248,"documentation":249,"maintenance":252,"popularity":259,"security":261,"testCoverage":262},{"hasLockfile":247},true,1778687006496,{"descriptionLength":250,"readmeSize":251},700,25712,{"closedIssues90d":253,"forks":254,"hasChangelog":255,"openIssues90d":256,"pushedAt":257,"stars":258},1,49,false,4,1778599393000,383,{"npmDownloads":260},51144,{"hasNpmPackage":247,"smitheryVerified":255},{"hasCi":247,"hasTests":247},{"updatedAt":264},1778687034478,{"basePath":266,"githubOwner":267,"githubRepo":227,"locale":18,"slug":268,"type":269},"skills/firecrawl-cli","firecrawl","firecrawl-cli","skill",{"_creationTime":271,"_id":272,"community":273,"display":274,"identity":277,"parentExtension":280,"providers":309,"relations":317,"tags":318,"workflow":319},1778686940775.5706,"k17axfavjpz72cd3qqzn86shb186ncqt",{"reviewCount":8},{"description":275,"installMethods":276,"name":267,"sourceUrl":14},"Scrape, search, crawl, and map the web with a single command.",{"claudeCode":267},{"basePath":278,"githubOwner":267,"githubRepo":227,"locale":18,"slug":227,"type":279},"","plugin",{"_creationTime":281,"_id":282,"community":283,"display":284,"identity":287,"providers":289,"relations":303,"tags":305,"workflow":306},1778686940775.5703,"k17be4khad1zr773wdg2m4yjhn86mjpr",{"reviewCount":8},{"description":285,"installMethods":286,"name":13,"sourceUrl":14},"CLI and Agent Skill for Firecrawl - Add scrape, search, and browsing capabilities to your AI agents",{"claudeCode":12},{"basePath":278,"githubOwner":267,"githubRepo":227,"locale":18,"slug":227,"type":288},"marketplace",{"evaluate":290,"extract":297},{"promptVersionExtension":291,"promptVersionScoring":220,"score":292,"tags":293,"targetMarket":231,"tier":296},"3.1.0",97,[226,227,294,230,228,295,229],"ai-agents","crawl","verified",{"commitSha":298,"license":299,"marketplace":300,"plugin":301},"HEAD","MIT",{"name":267,"pluginCount":253},{"mcpCount":8,"provider":302,"skillCount":8},"classify",{"repoId":304},"kd7csd1wb06dg9c1jfy5063f2586ne60",[294,229,227,295,230,228,226],{"evaluatedAt":307,"extractAt":308,"updatedAt":307},1778686962954,1778686940775,{"evaluate":310,"extract":314},{"promptVersionExtension":219,"promptVersionScoring":220,"score":311,"tags":312,"targetMarket":231,"tier":232},78,[226,227,229,230,228,295,313],"agent",{"commitSha":298,"plugin":315},{"mcpCount":8,"provider":302,"skillCount":316},9,{"parentExtensionId":282,"repoId":304},[313,229,227,295,230,228,226],{"evaluatedAt":320,"extractAt":308,"updatedAt":320},1778686985594,{"evaluate":322,"extract":324},{"promptVersionExtension":219,"promptVersionScoring":220,"score":223,"tags":323,"targetMarket":231,"tier":232},[226,227,228,229,230],{"commitSha":298,"license":299},{"parentExtensionId":272,"repoId":304},{"_creationTime":327,"_id":304,"identity":328,"providers":329,"workflow":504},1778686934511.6108,{"githubOwner":267,"githubRepo":227,"sourceUrl":14},{"classify":330,"discover":496,"extract":499,"github":500,"npm":503},{"commitSha":298,"extensions":331},[332,342,368,377,382,390,398,406,414,422,430,438],{"basePath":278,"displayName":267,"installMethods":333,"rationale":334,"selectedPaths":335,"source":341,"sourceLanguage":18,"type":288},{"claudeCode":12},"marketplace.json at .claude-plugin/marketplace.json",[336,339],{"path":337,"priority":338},".claude-plugin/marketplace.json","mandatory",{"path":340,"priority":338},"README.md","rule",{"basePath":278,"description":275,"displayName":267,"installMethods":343,"rationale":344,"selectedPaths":345,"source":341,"sourceLanguage":18,"type":279},{"claudeCode":267},"plugin manifest at .claude-plugin/plugin.json",[346,348,349,352,354,356,358,360,362,364,366],{"path":347,"priority":338},".claude-plugin/plugin.json",{"path":340,"priority":338},{"path":350,"priority":351},"skills/firecrawl-agent/SKILL.md","medium",{"path":353,"priority":351},"skills/firecrawl-cli/SKILL.md",{"path":355,"priority":351},"skills/firecrawl-crawl/SKILL.md",{"path":357,"priority":351},"skills/firecrawl-download/SKILL.md",{"path":359,"priority":351},"skills/firecrawl-interact/SKILL.md",{"path":361,"priority":351},"skills/firecrawl-map/SKILL.md",{"path":363,"priority":351},"skills/firecrawl-parse/SKILL.md",{"path":365,"priority":351},"skills/firecrawl-scrape/SKILL.md",{"path":367,"priority":351},"skills/firecrawl-search/SKILL.md",{"basePath":369,"description":370,"displayName":371,"installMethods":372,"rationale":373,"selectedPaths":374,"source":341,"sourceLanguage":18,"type":269},"skills/firecrawl-agent","AI-powered autonomous data extraction that navigates complex sites and returns structured JSON. Use this skill when the user wants structured data from websites, needs to extract pricing tiers, product listings, directory entries, or any data as JSON with a schema. Triggers on \"extract structured data\", \"get all the products\", \"pull pricing info\", \"extract as JSON\", or when the user provides a JSON schema for website data. More powerful than simple scraping for multi-page structured extraction.\n","firecrawl-agent",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-agent/SKILL.md",[375],{"path":376,"priority":338},"SKILL.md",{"basePath":266,"description":10,"displayName":267,"installMethods":378,"rationale":379,"selectedPaths":380,"source":341,"sourceLanguage":18,"type":269},{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-cli/SKILL.md",[381],{"path":376,"priority":338},{"basePath":383,"description":384,"displayName":385,"installMethods":386,"rationale":387,"selectedPaths":388,"source":341,"sourceLanguage":18,"type":269},"skills/firecrawl-crawl","Bulk extract content from an entire website or site section. Use this skill when the user wants to crawl a site, extract all pages from a docs section, bulk-scrape multiple pages following links, or says \"crawl\", \"get all the pages\", \"extract everything under /docs\", \"bulk extract\", or needs content from many pages on the same site. Handles depth limits, path filtering, and concurrent extraction.\n","firecrawl-crawl",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-crawl/SKILL.md",[389],{"path":376,"priority":338},{"basePath":391,"description":392,"displayName":393,"installMethods":394,"rationale":395,"selectedPaths":396,"source":341,"sourceLanguage":18,"type":269},"skills/firecrawl-download","Download an entire website as local files — markdown, screenshots, or multiple formats per page. Use this skill when the user wants to save a site locally, download documentation for offline use, bulk-save pages as files, or says \"download the site\", \"save as local files\", \"offline copy\", \"download all the docs\", or \"save for reference\". Combines site mapping and scraping into organized local directories.\n","firecrawl-download",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-download/SKILL.md",[397],{"path":376,"priority":338},{"basePath":399,"description":400,"displayName":401,"installMethods":402,"rationale":403,"selectedPaths":404,"source":341,"sourceLanguage":18,"type":269},"skills/firecrawl-interact","Control and interact with a live browser session on any scraped page — click buttons, fill forms, navigate flows, and extract data using natural language prompts or code. Use when the user needs to interact with a webpage beyond simple scraping: logging into a site, submitting forms, clicking through pagination, handling infinite scroll, navigating multi-step checkout or wizard flows, or when a regular scrape failed because content is behind JavaScript interaction. Also useful for authenticated scraping via profiles. Triggers on \"interact\", \"click\", \"fill out the form\", \"log in to\", \"sign in\", \"submit\", \"paginated\", \"next page\", \"infinite scroll\", \"interact with the page\", \"navigate to\", \"open a session\", or \"scrape failed\".\n","firecrawl-interact",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-interact/SKILL.md",[405],{"path":376,"priority":338},{"basePath":407,"description":408,"displayName":409,"installMethods":410,"rationale":411,"selectedPaths":412,"source":341,"sourceLanguage":18,"type":269},"skills/firecrawl-map","Discover and list all URLs on a website, with optional search filtering. Use this skill when the user wants to find a specific page on a large site, list all URLs, see the site structure, find where something is on a domain, or says \"map the site\", \"find the URL for\", \"what pages are on\", or \"list all pages\". Essential when the user knows which site but not which exact page.\n","firecrawl-map",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-map/SKILL.md",[413],{"path":376,"priority":338},{"basePath":415,"description":416,"displayName":417,"installMethods":418,"rationale":419,"selectedPaths":420,"source":341,"sourceLanguage":18,"type":269},"skills/firecrawl-parse","Efficiently extract and convert the contents of any local file—such as PDF, DOCX, DOC, ODT, RTF, XLSX, XLS, or HTML—into clean, well-formatted markdown saved to disk. Use this skill whenever the user requests to parse, read, or extract information from a file on their computer, including phrases like “parse this PDF”, “convert this document”, “read this file”, “extract text from”, or when a local file path (not a URL) is provided. This skill offers advanced options like generating AI-powered summaries and answering questions based on the file's content. Prefer this tool over `scrape` when handling local files to deliver precise, structured outputs for downstream tasks.\n","firecrawl-parse",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-parse/SKILL.md",[421],{"path":376,"priority":338},{"basePath":423,"description":424,"displayName":425,"installMethods":426,"rationale":427,"selectedPaths":428,"source":341,"sourceLanguage":18,"type":269},"skills/firecrawl-scrape","Extract clean markdown from any URL, including JavaScript-rendered SPAs. Use this skill whenever the user provides a URL and wants its content, says \"scrape\", \"grab\", \"fetch\", \"pull\", \"get the page\", \"extract from this URL\", or \"read this webpage\". Handles JS-rendered pages, multiple concurrent URLs, and returns LLM-optimized markdown. Use this instead of WebFetch for any webpage content extraction.\n","firecrawl-scrape",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-scrape/SKILL.md",[429],{"path":376,"priority":338},{"basePath":431,"description":432,"displayName":433,"installMethods":434,"rationale":435,"selectedPaths":436,"source":341,"sourceLanguage":18,"type":269},"skills/firecrawl-search","Web search with full page content extraction. Use this skill whenever the user asks to search the web, find articles, research a topic, look something up, find recent news, discover sources, or says \"search for\", \"find me\", \"look up\", \"what are people saying about\", or \"find articles about\". Returns real search results with optional full-page markdown — not just snippets. Provides capabilities beyond Claude's built-in WebSearch.\n","firecrawl-search",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-search/SKILL.md",[437],{"path":376,"priority":338},{"basePath":278,"description":439,"displayName":268,"installMethods":440,"license":441,"rationale":442,"selectedPaths":443,"source":341,"sourceLanguage":18,"type":227},"Command-line interface for Firecrawl. Scrape, crawl, and extract data from any website directly from your terminal.",{"npm":268},"ISC","cli ecosystem detected at /",[444,446,447,450,452,454,456,458,460,462,464,466,468,470,472,474,476,478,480,482,484,486,488,490,492,494],{"path":445,"priority":338},"package.json",{"path":340,"priority":338},{"path":448,"priority":449},"src/index.ts","low",{"path":451,"priority":449},"src/commands/agent.ts",{"path":453,"priority":449},"src/commands/browser.ts",{"path":455,"priority":449},"src/commands/config.ts",{"path":457,"priority":449},"src/commands/crawl.ts",{"path":459,"priority":449},"src/commands/create.ts",{"path":461,"priority":449},"src/commands/credit-usage.ts",{"path":463,"priority":449},"src/commands/env.ts",{"path":465,"priority":449},"src/commands/experimental/backends.ts",{"path":467,"priority":449},"src/commands/experimental/index.ts",{"path":469,"priority":449},"src/commands/experimental/shared.ts",{"path":471,"priority":449},"src/commands/init.ts",{"path":473,"priority":449},"src/commands/interact.ts",{"path":475,"priority":449},"src/commands/login.ts",{"path":477,"priority":449},"src/commands/logout.ts",{"path":479,"priority":449},"src/commands/map.ts",{"path":481,"priority":449},"src/commands/parse.ts",{"path":483,"priority":449},"src/commands/scrape.ts",{"path":485,"priority":449},"src/commands/search.ts",{"path":487,"priority":449},"src/commands/setup.ts",{"path":489,"priority":449},"src/commands/skills-install.ts",{"path":491,"priority":449},"src/commands/skills-native.ts",{"path":493,"priority":449},"src/commands/status.ts",{"path":495,"priority":449},"src/commands/version.ts",{"sources":497},[498],"manual",{"npmPackage":268},{"closedIssues90d":253,"description":285,"forks":254,"homepage":501,"openIssues90d":256,"pushedAt":257,"readmeSize":251,"stars":258,"topics":502},"http://docs.firecrawl.dev/cli",[],{"downloads":260},{"classifiedAt":505,"discoverAt":506,"extractAt":507,"githubAt":507,"npmAt":508,"updatedAt":505},1778686940560,1778686934511,1778686936677,1778686938425,[229,227,230,228,226],{"evaluatedAt":264,"extractAt":308,"updatedAt":264},[],[513,541,572,594,620,640],{"_creationTime":514,"_id":515,"community":516,"display":517,"identity":523,"providers":527,"relations":534,"tags":536,"workflow":537},1778697205743.894,"k17ej9360g7468qswf3svnrx9186m1ng",{"reviewCount":8},{"description":518,"installMethods":519,"name":521,"sourceUrl":522},"Browser automation CLI for AI agents. Use when the user needs to interact with websites, including navigating pages, filling forms, clicking buttons, taking screenshots, extracting data, testing web apps, or automating any browser task. Triggers include requests to \"open a website\", \"fill out a form\", \"click a button\", \"take a screenshot\", \"scrape data from a page\", \"test this web app\", \"login to a site\", \"automate browser actions\", or any task requiring programmatic web interaction.",{"claudeCode":520},"shanraisshan/claude-code-best-practice","agent-browser","https://github.com/shanraisshan/claude-code-best-practice",{"basePath":524,"githubOwner":525,"githubRepo":526,"locale":18,"slug":521,"type":269},".claude/skills/agent-browser","shanraisshan","claude-code-best-practice",{"evaluate":528,"extract":533},{"promptVersionExtension":219,"promptVersionScoring":220,"score":529,"tags":530,"targetMarket":231,"tier":296},100,[531,226,532,227,229],"browser-automation","testing",{"commitSha":298},{"repoId":535},"kd74710g49kxgwbfjxeb7s132d86myxr",[229,531,227,532,226],{"evaluatedAt":538,"extractAt":539,"updatedAt":540},1778697222051,1778697205743,1778697438718,{"_creationTime":542,"_id":543,"community":544,"display":545,"identity":551,"providers":556,"relations":565,"tags":568,"workflow":569},1778699234184.6157,"k177tdbfgqmwhtaqv771f2ych586nne9",{"reviewCount":8},{"description":546,"installMethods":547,"name":549,"sourceUrl":550},"Worktree-first dev environment manager for issues, PRs, and features with optional tmux sessions",{"claudeCode":548},"Yeachan-Heo/oh-my-claudecode","Project Session Manager","https://github.com/Yeachan-Heo/oh-my-claudecode",{"basePath":552,"githubOwner":553,"githubRepo":554,"locale":18,"slug":555,"type":269},"skills/project-session-manager","Yeachan-Heo","oh-my-claudecode","project-session-manager",{"evaluate":557,"extract":564},{"promptVersionExtension":219,"promptVersionScoring":220,"score":529,"tags":558,"targetMarket":231,"tier":296},[559,560,561,562,229,227,563],"git","development-environment","workflow","tmux","developer-tool",{"commitSha":298,"license":299},{"parentExtensionId":566,"repoId":567},"k17brg5egdw1jbncj1j4wfv3fh86n639","kd74zv63fryf9prygtq7gf4es986n22y",[229,227,563,560,559,562,561],{"evaluatedAt":570,"extractAt":571,"updatedAt":570},1778699613343,1778699234184,{"_creationTime":573,"_id":574,"community":575,"display":576,"identity":579,"providers":582,"relations":590,"tags":591,"workflow":592},1778699234184.611,"k179b6dkc777g1rgyecze04wqn86m6y4",{"reviewCount":8},{"description":577,"installMethods":578,"name":554,"sourceUrl":550},"Process-first advisor routing for Claude, Codex, or Gemini via `omc ask`, with artifact capture and no raw CLI assembly",{"claudeCode":548},{"basePath":580,"githubOwner":553,"githubRepo":554,"locale":18,"slug":581,"type":269},"skills/ask","ask",{"evaluate":583,"extract":589},{"promptVersionExtension":219,"promptVersionScoring":220,"score":529,"tags":584,"targetMarket":231,"tier":296},[227,229,585,586,587,588],"llm","prompting","code-review","artifact-generation",{"commitSha":298,"license":299},{"parentExtensionId":566,"repoId":567},[588,229,227,587,585,586],{"evaluatedAt":593,"extractAt":571,"updatedAt":593},1778699303045,{"_creationTime":595,"_id":596,"community":597,"display":598,"identity":604,"providers":608,"relations":614,"tags":616,"workflow":617},1778697652123.8928,"k171pew5empzzrfghyg9nqrk6n86nqa9",{"reviewCount":8},{"description":599,"installMethods":600,"name":602,"sourceUrl":603},"Use gh for GitHub issues, PR status, CI/logs, comments, reviews, releases, and API queries.",{"claudeCode":601},"steipete/clawdis","github","https://github.com/steipete/clawdis",{"basePath":605,"githubOwner":606,"githubRepo":607,"locale":18,"slug":602,"type":269},"skills/github","steipete","clawdis",{"evaluate":609,"extract":613},{"promptVersionExtension":219,"promptVersionScoring":220,"score":529,"tags":610,"targetMarket":231,"tier":296},[602,227,611,612,229],"api","developer-tools",{"commitSha":298},{"repoId":615},"kd738npxg9yh3xf3vddzy9fyfh86nhng",[611,229,227,612,602],{"evaluatedAt":618,"extractAt":619,"updatedAt":618},1778698569289,1778697652123,{"_creationTime":621,"_id":622,"community":623,"display":624,"identity":628,"providers":630,"relations":636,"tags":637,"workflow":638},1778697652123.8857,"k179k1n5nw0md1x1j3dak31evn86mb5f",{"reviewCount":8},{"description":625,"installMethods":626,"name":627,"sourceUrl":603},"Use when reviewing, reproducing, or proving OpenClaw Telegram behavior with a real Telegram user on Crabbox, including PR review workflows that need an agent-controlled Telegram Desktop recording, TDLib user-driver commands, Convex-leased credentials, WebVNC observation, and motion-trimmed artifacts.",{"claudeCode":601},"telegram-crabbox-e2e-proof",{"basePath":629,"githubOwner":606,"githubRepo":607,"locale":18,"slug":627,"type":269},".agents/skills/telegram-crabbox-e2e-proof",{"evaluate":631,"extract":635},{"promptVersionExtension":219,"promptVersionScoring":220,"score":529,"tags":632,"targetMarket":231,"tier":296},[532,633,229,634,227],"telegram","qa",{"commitSha":298},{"repoId":615},[229,227,634,633,532],{"evaluatedAt":639,"extractAt":619,"updatedAt":639},1778698038113,{"_creationTime":641,"_id":642,"community":643,"display":644,"identity":650,"providers":654,"relations":660,"tags":662,"workflow":663},1778696993586.7058,"k174ty9vbfgj7cv3v2szdwjgh986mrsf",{"reviewCount":8},{"description":645,"installMethods":646,"name":648,"sourceUrl":649},"Search ChatGPT and extract the full response + hydration JSON that powers the UI.\nAttaches to a running Chrome instance (port 9222 by default), opens ChatGPT,\nsubmits a query, waits for the streamed response, and returns structured data:\nmessages, product cards, hydration JSON, and API calls.\nUse when asked to \"search chatgpt\", \"ask chatgpt\", \"chatgpt search\",\n\"get chatgpt response\", or \"scrape chatgpt\".\n",{"claudeCode":647},"SeifBenayed/claude-code-sdk","chatgpt-search","https://github.com/SeifBenayed/claude-code-sdk",{"basePath":651,"githubOwner":652,"githubRepo":653,"locale":18,"slug":648,"type":269},".claude/skills/chatgpt-search","SeifBenayed","claude-code-sdk",{"evaluate":655,"extract":659},{"promptVersionExtension":219,"promptVersionScoring":220,"score":529,"tags":656,"targetMarket":231,"tier":296},[657,658,531,230,226],"chatgpt","scraping",{"commitSha":298},{"repoId":661},"kd78s53c1852h5p7c3qem663xs86njab",[531,657,230,658,226],{"evaluatedAt":664,"extractAt":665,"updatedAt":664},1778697026743,1778696993586]