[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"extension-skill-firecrawl-firecrawl-scrape-zh-CN":3,"guides-for-firecrawl-firecrawl-scrape":505,"similar-k17chm1rbb9bh8xn1d9xbgyzwd86ms24-zh-CN":506},{"_creationTime":4,"_id":5,"children":6,"community":7,"display":9,"evaluation":15,"identity":257,"isFallback":239,"parentExtension":262,"providers":315,"relations":319,"repo":320,"tags":503,"workflow":504},1778686940775.5725,"k17chm1rbb9bh8xn1d9xbgyzwd86ms24",[],{"reviewCount":8},0,{"description":10,"installMethods":11,"name":13,"sourceUrl":14},"Extract clean markdown from any URL, including JavaScript-rendered SPAs. Use this skill whenever the user provides a URL and wants its content, says \"scrape\", \"grab\", \"fetch\", \"pull\", \"get the page\", \"extract from this URL\", or \"read this webpage\". Handles JS-rendered pages, multiple concurrent URLs, and returns LLM-optimized markdown. Use this instead of WebFetch for any webpage content extraction.\n",{"claudeCode":12},"firecrawl/cli","Firecrawl CLI","https://github.com/firecrawl/cli",{"_creationTime":16,"_id":17,"extensionId":5,"locale":18,"result":19,"trustSignals":237,"workflow":255},1778687209738.6738,"kn79macrhc20z2v09hcbtsscnx86mb21","en",{"checks":20,"evaluatedAt":193,"extensionSummary":194,"features":195,"nonGoals":203,"practices":207,"prerequisites":208,"promptVersionExtension":211,"promptVersionScoring":212,"purpose":213,"rationale":214,"score":215,"summary":216,"tags":217,"targetMarket":225,"tier":226,"useCases":227,"workflow":232},[21,26,29,32,36,39,43,47,50,53,57,61,64,68,71,74,77,80,83,86,90,94,99,103,107,110,114,117,121,124,127,130,133,136,139,143,147,150,154,158,161,164,167,170,174,177,180,183,186,190],{"category":22,"check":23,"severity":24,"summary":25},"Practical Utility","Problem relevance","pass","The description clearly states the problem of extracting clean markdown from URLs, including JS-rendered SPAs, and identifies specific user intents and scenarios for its use.",{"category":22,"check":27,"severity":24,"summary":28},"Unique selling proposition","The skill offers significant value beyond a basic prompt by handling JS-rendered pages, concurrent URLs, and providing LLM-optimized markdown, differentiating itself from generic web fetching.",{"category":22,"check":30,"severity":24,"summary":31},"Production readiness","The skill is production-ready, offering comprehensive functionality for web page content extraction from simple scrapes to interactive sessions and bulk crawling, covering the complete lifecycle.",{"category":33,"check":34,"severity":24,"summary":35},"Scope","Single responsibility principle","The extension focuses on web content extraction and interaction, with distinct commands for search, scrape, map, crawl, and agent tasks, maintaining a coherent scope.",{"category":33,"check":37,"severity":24,"summary":38},"Description quality","The displayed description accurately reflects the skill's capabilities, including handling JS-rendered pages and concurrent URLs, and provides clear usage instructions.",{"category":40,"check":41,"severity":24,"summary":42},"Invocation","Scoped tools","The CLI exposes a set of narrowly scoped tools like `scrape`, `search`, `crawl`, etc., facilitating precise agent selection.",{"category":44,"check":45,"severity":24,"summary":46},"Documentation","Configuration & parameter reference","All options and parameters for commands like `scrape`, `search`, and `crawl` are clearly documented in the README, including formats, limits, and timeouts.",{"category":33,"check":48,"severity":24,"summary":49},"Tool naming","Tool names such as `scrape`, `search`, `crawl`, and `agent` are descriptive and clearly indicate their function.",{"category":33,"check":51,"severity":24,"summary":52},"Minimal I/O surface","Tool parameters are specific and well-defined (e.g., `--format`, `--limit`), and outputs are structured (JSON, markdown) or file-based, avoiding unnecessary data.",{"category":54,"check":55,"severity":24,"summary":56},"License","License usability","The project uses the MIT license, which is permissive and allows for broad usability.",{"category":58,"check":59,"severity":24,"summary":60},"Maintenance","Commit recency","The latest commit was on May 12, 2026, indicating recent maintenance activity.",{"category":58,"check":62,"severity":24,"summary":63},"Dependency Management","The project uses npm and includes a lockfile, indicating good dependency management practices.",{"category":65,"check":66,"severity":24,"summary":67},"Security","Secret Management","The CLI handles API keys securely, allowing them via environment variables or per-command flags, and does not echo them to stdout.",{"category":65,"check":69,"severity":24,"summary":70},"Injection","The skill processes URLs and user prompts as data, and the use of specific commands like `scrape` and `search` with defined parameters mitigates injection risks.",{"category":65,"check":72,"severity":24,"summary":73},"Transitive Supply-Chain Grenades","The CLI appears to bundle dependencies and execute commands locally, avoiding runtime downloads of uncommitted code or remote execution of arbitrary scripts.",{"category":65,"check":75,"severity":24,"summary":76},"Sandbox Isolation","The CLI operates as a command-line tool, and its operations like scraping and crawling are confined to network requests and local file output within specified paths.",{"category":65,"check":78,"severity":24,"summary":79},"Sandbox escape primitives","No evidence of detached process spawns or retry loops around denied tool calls was found in the CLI's structure.",{"category":65,"check":81,"severity":24,"summary":82},"Data Exfiltration","The CLI's telemetry is opt-in and explicitly states that command data, URLs, or file contents are not collected. Outbound calls are primarily for API requests to Firecrawl services.",{"category":65,"check":84,"severity":24,"summary":85},"Hidden Text Tricks","The bundled documentation and code do not contain hidden text tricks, Unicode control characters, or other obfuscation methods.",{"category":87,"check":88,"severity":24,"summary":89},"Hooks","Opaque code execution","The CLI is composed of standard JavaScript code and not obfuscated via base64 payloads or runtime script fetching.",{"category":91,"check":92,"severity":24,"summary":93},"Portability","Structural Assumption","The CLI's commands operate based on provided URLs and parameters, and its output handling (saving to specified paths) does not make rigid assumptions about the user's project structure.",{"category":95,"check":96,"severity":97,"summary":98},"Trust","Issues Attention","warning","In the last 90 days, 4 issues were opened and 1 was closed, indicating a low closure rate and potentially slow maintainer response.",{"category":100,"check":101,"severity":24,"summary":102},"Versioning","Release Management","The CLI declares a version (v1.16.2) in its README and package installation, facilitating version tracking.",{"category":104,"check":105,"severity":24,"summary":106},"Execution","Validation","The CLI utilizes command-line argument parsing, and specific options like `--format`, `--limit`, and `--schema` suggest structured input handling.",{"category":65,"check":108,"severity":24,"summary":109},"Unguarded Destructive Operations","The CLI's primary operations are data retrieval and interaction, with no inherently destructive commands like file deletion or system modification without explicit user action.",{"category":111,"check":112,"severity":24,"summary":113},"Code Execution","Error Handling","The CLI provides user-friendly error messages and status updates, with clear options for handling different scenarios like authentication or rate limiting.",{"category":111,"check":115,"severity":24,"summary":116},"Logging","The CLI provides status updates and output redirection, and its `--status` command offers a form of audit.",{"category":118,"check":119,"severity":24,"summary":120},"Compliance","GDPR","The CLI focuses on web content extraction and does not appear to operate on personal data beyond what is publicly available on websites, with opt-in telemetry.",{"category":118,"check":122,"severity":24,"summary":123},"Target market","The extension is designed for general web scraping and data extraction and does not exhibit any regional or jurisdictional limitations.",{"category":91,"check":125,"severity":24,"summary":126},"Runtime stability","The CLI is a Node.js application installed via npm, making it generally portable across POSIX-compliant systems and Windows.",{"category":44,"check":128,"severity":24,"summary":129},"README","The README is comprehensive, detailing installation, authentication, commands, options, and examples.",{"category":33,"check":131,"severity":24,"summary":132},"Tool surface size","The CLI exposes a focused set of 7 primary commands (`scrape`, `search`, `map`, `crawl`, `credit-usage`, `agent`, `interact`, `config`, `view-config`, `login`, `logout`), which is within the ideal range.",{"category":40,"check":134,"severity":24,"summary":135},"Overlapping near-synonym tools","The tools provided (`scrape`, `search`, `crawl`, `agent`, `interact`) have distinct functions and do not exhibit significant overlap.",{"category":44,"check":137,"severity":24,"summary":138},"Phantom features","All features advertised in the README and SKILL.md (e.g., `scrape`, `search`, `crawl`, `agent`) have corresponding implemented commands and functionalities.",{"category":140,"check":141,"severity":24,"summary":142},"Install","Installation instruction","The README provides clear installation instructions via npm and includes multiple copy-pasteable invocation examples for various commands.",{"category":144,"check":145,"severity":24,"summary":146},"Errors","Actionable error messages","The CLI provides informative error messages and status updates, guiding users on authentication, rate limits, and command usage.",{"category":104,"check":148,"severity":24,"summary":149},"Pinned dependencies","The use of `npm install -g firecrawl-cli` and the presence of a lockfile indicate that dependencies are managed and pinned.",{"category":33,"check":151,"severity":152,"summary":153},"Dry-run preview","not_applicable","The primary operations of the CLI are data retrieval and interaction, which do not involve destructive operations requiring a dry-run mode.",{"category":155,"check":156,"severity":24,"summary":157},"Protocol","Idempotent retry & timeouts","The CLI handles network requests with timeouts and provides status updates for ongoing jobs, indicating good protocol practices.",{"category":118,"check":159,"severity":24,"summary":160},"Telemetry opt-in","Telemetry is strictly opt-in via an environment variable (`FIRECRAWL_NO_TELEMETRY=1`) and does not collect sensitive user data.",{"category":40,"check":162,"severity":24,"summary":163},"Precise Purpose","The SKILL.md and README clearly define the purpose of extracting markdown from URLs and when to use it, including specific trigger phrases and boundaries.",{"category":40,"check":165,"severity":24,"summary":166},"Concise Frontmatter","The SKILL.md frontmatter is concise and effectively summarizes the core capability and trigger phrases for the `firecrawl-scrape` skill.",{"category":44,"check":168,"severity":24,"summary":169},"Concise Body","The README.md and SKILL.md files are well-structured and avoid excessive length, delegating detailed information to specific command sections and examples.",{"category":171,"check":172,"severity":24,"summary":173},"Context","Progressive Disclosure","The README provides a high-level overview with detailed explanations for each command, effectively using progressive disclosure.",{"category":171,"check":175,"severity":152,"summary":176},"Forked exploration","This extension is a CLI tool, not a deep exploration skill requiring a forked context.",{"category":22,"check":178,"severity":24,"summary":179},"Usage examples","The README provides numerous end-to-end, ready-to-use examples for each command, demonstrating input, invocation, and expected output.",{"category":22,"check":181,"severity":24,"summary":182},"Edge cases","The documentation addresses various options and scenarios, such as waiting for JS rendering, handling multiple formats, and specifying output paths, implicitly covering some edge cases.",{"category":111,"check":184,"severity":152,"summary":185},"Tool Fallback","The extension is a standalone CLI tool and does not rely on external MCP servers or other tools that would require fallbacks.",{"category":187,"check":188,"severity":24,"summary":189},"Safety","Halt on unexpected state","The CLI provides clear error messages for issues like authentication failures or invalid inputs, halting execution gracefully.",{"category":91,"check":191,"severity":24,"summary":192},"Cross-skill coupling","The Firecrawl CLI commands are self-contained and do not implicitly rely on other skills being loaded or present.",1778687208974,"This is a command-line interface (CLI) tool that allows users to search, scrape, interact with, crawl, and map websites. It can extract markdown, HTML, links, and images from URLs, including JavaScript-rendered single-page applications (SPAs). It also supports AI-powered agent tasks for complex data extraction and interactive browser sessions.",[196,197,198,199,200,201,202],"Extract clean markdown from any URL","Handle JavaScript-rendered SPAs","Perform concurrent URL scraping","Search the web and scrape results","Crawl entire websites","Interact with scraped pages via prompts or code","Run AI agent tasks for autonomous data extraction",[204,205,206],"Replacing browser functionality for end-users.","Performing real-time interaction beyond controlled sessions.","Executing arbitrary code on remote servers (beyond user-provided scripts in isolated contexts).",[],[209,210],"Node.js and npm installed","A Firecrawl API key (optional for self-hosted instances)","3.0.0","4.4.0","To provide a versatile command-line tool for extracting and interacting with web content, especially for developers and automated workflows.","The extension is robust, well-documented, and production-ready with a clear scope and minimal security risks. The only minor finding is the low issue closure rate.",95,"A powerful and well-documented CLI tool for extracting web content.",[218,219,220,221,222,223,224],"web-scraping","markdown","data-extraction","cli","spa","javascript-rendering","content-retrieval","global","community",[228,229,230,231],"Extracting article content from blog posts for summarization or analysis.","Scraping product details and prices from e-commerce sites.","Gathering research data from multiple web pages concurrently.","Automating website audits or content discovery.",[233,234,235,236],"Define the task (e.g., scrape a URL, search for information).","Invoke the appropriate Firecrawl CLI command with relevant options.","Process the output (e.g., save to file, pipe to another command, analyze JSON).","For complex tasks, use the agent or interact commands for guided or autonomous web interaction.",{"codeQuality":238,"collectedAt":240,"documentation":241,"maintenance":244,"popularity":251,"security":253,"testCoverage":254},{"hasLockfile":239},true,1778687185008,{"descriptionLength":242,"readmeSize":243},403,25712,{"closedIssues90d":245,"forks":246,"hasChangelog":247,"openIssues90d":248,"pushedAt":249,"stars":250},1,49,false,4,1778599393000,383,{"npmDownloads":252},51144,{"hasNpmPackage":239,"smitheryVerified":247},{"hasCi":239,"hasTests":239},{"updatedAt":256},1778687209738,{"basePath":258,"githubOwner":259,"githubRepo":221,"locale":18,"slug":260,"type":261},"skills/firecrawl-scrape","firecrawl","firecrawl-scrape","skill",{"_creationTime":263,"_id":264,"community":265,"display":266,"identity":269,"parentExtension":272,"providers":303,"relations":311,"tags":312,"workflow":313},1778686940775.5706,"k17axfavjpz72cd3qqzn86shb186ncqt",{"reviewCount":8},{"description":267,"installMethods":268,"name":259,"sourceUrl":14},"Scrape, search, crawl, and map the web with a single command.",{"claudeCode":259},{"basePath":270,"githubOwner":259,"githubRepo":221,"locale":18,"slug":221,"type":271},"","plugin",{"_creationTime":273,"_id":274,"community":275,"display":276,"identity":279,"providers":281,"relations":297,"tags":299,"workflow":300},1778686940775.5703,"k17be4khad1zr773wdg2m4yjhn86mjpr",{"reviewCount":8},{"description":277,"installMethods":278,"name":13,"sourceUrl":14},"CLI and Agent Skill for Firecrawl - Add scrape, search, and browsing capabilities to your AI agents",{"claudeCode":12},{"basePath":270,"githubOwner":259,"githubRepo":221,"locale":18,"slug":221,"type":280},"marketplace",{"evaluate":282,"extract":291},{"promptVersionExtension":283,"promptVersionScoring":212,"score":284,"tags":285,"targetMarket":225,"tier":290},"3.1.0",97,[218,221,286,220,287,288,289],"ai-agents","search","crawl","automation","verified",{"commitSha":292,"license":293,"marketplace":294,"plugin":295},"HEAD","MIT",{"name":259,"pluginCount":245},{"mcpCount":8,"provider":296,"skillCount":8},"classify",{"repoId":298},"kd7csd1wb06dg9c1jfy5063f2586ne60",[286,289,221,288,220,287,218],{"evaluatedAt":301,"extractAt":302,"updatedAt":301},1778686962954,1778686940775,{"evaluate":304,"extract":308},{"promptVersionExtension":211,"promptVersionScoring":212,"score":305,"tags":306,"targetMarket":225,"tier":226},78,[218,221,289,220,287,288,307],"agent",{"commitSha":292,"plugin":309},{"mcpCount":8,"provider":296,"skillCount":310},9,{"parentExtensionId":274,"repoId":298},[307,289,221,288,220,287,218],{"evaluatedAt":314,"extractAt":302,"updatedAt":314},1778686985594,{"evaluate":316,"extract":318},{"promptVersionExtension":211,"promptVersionScoring":212,"score":215,"tags":317,"targetMarket":225,"tier":226},[218,219,220,221,222,223,224],{"commitSha":292,"license":293},{"parentExtensionId":264,"repoId":298},{"_creationTime":321,"_id":298,"identity":322,"providers":323,"workflow":498},1778686934511.6108,{"githubOwner":259,"githubRepo":221,"sourceUrl":14},{"classify":324,"discover":490,"extract":493,"github":494,"npm":497},{"commitSha":292,"extensions":325},[326,336,362,371,378,386,394,402,410,418,423,431],{"basePath":270,"displayName":259,"installMethods":327,"rationale":328,"selectedPaths":329,"source":335,"sourceLanguage":18,"type":280},{"claudeCode":12},"marketplace.json at .claude-plugin/marketplace.json",[330,333],{"path":331,"priority":332},".claude-plugin/marketplace.json","mandatory",{"path":334,"priority":332},"README.md","rule",{"basePath":270,"description":267,"displayName":259,"installMethods":337,"rationale":338,"selectedPaths":339,"source":335,"sourceLanguage":18,"type":271},{"claudeCode":259},"plugin manifest at .claude-plugin/plugin.json",[340,342,343,346,348,350,352,354,356,358,360],{"path":341,"priority":332},".claude-plugin/plugin.json",{"path":334,"priority":332},{"path":344,"priority":345},"skills/firecrawl-agent/SKILL.md","medium",{"path":347,"priority":345},"skills/firecrawl-cli/SKILL.md",{"path":349,"priority":345},"skills/firecrawl-crawl/SKILL.md",{"path":351,"priority":345},"skills/firecrawl-download/SKILL.md",{"path":353,"priority":345},"skills/firecrawl-interact/SKILL.md",{"path":355,"priority":345},"skills/firecrawl-map/SKILL.md",{"path":357,"priority":345},"skills/firecrawl-parse/SKILL.md",{"path":359,"priority":345},"skills/firecrawl-scrape/SKILL.md",{"path":361,"priority":345},"skills/firecrawl-search/SKILL.md",{"basePath":363,"description":364,"displayName":365,"installMethods":366,"rationale":367,"selectedPaths":368,"source":335,"sourceLanguage":18,"type":261},"skills/firecrawl-agent","AI-powered autonomous data extraction that navigates complex sites and returns structured JSON. Use this skill when the user wants structured data from websites, needs to extract pricing tiers, product listings, directory entries, or any data as JSON with a schema. Triggers on \"extract structured data\", \"get all the products\", \"pull pricing info\", \"extract as JSON\", or when the user provides a JSON schema for website data. More powerful than simple scraping for multi-page structured extraction.\n","firecrawl-agent",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-agent/SKILL.md",[369],{"path":370,"priority":332},"SKILL.md",{"basePath":372,"description":373,"displayName":259,"installMethods":374,"rationale":375,"selectedPaths":376,"source":335,"sourceLanguage":18,"type":261},"skills/firecrawl-cli","Search, scrape, and interact with the web via the Firecrawl CLI. Use this skill whenever the user wants to search the web, find articles, research a topic, look something up online, scrape a webpage, grab content from a URL, get data from a website, crawl documentation, download a site, or interact with pages that need clicks or logins. Also use when they say \"fetch this page\", \"pull the content from\", \"get the page at https://\", or reference external websites. This provides real-time web search with full page content and interact capabilities — beyond what Claude can do natively with built-in tools. Do NOT trigger for local file operations, git commands, deployments, or code editing tasks.\n",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-cli/SKILL.md",[377],{"path":370,"priority":332},{"basePath":379,"description":380,"displayName":381,"installMethods":382,"rationale":383,"selectedPaths":384,"source":335,"sourceLanguage":18,"type":261},"skills/firecrawl-crawl","Bulk extract content from an entire website or site section. Use this skill when the user wants to crawl a site, extract all pages from a docs section, bulk-scrape multiple pages following links, or says \"crawl\", \"get all the pages\", \"extract everything under /docs\", \"bulk extract\", or needs content from many pages on the same site. Handles depth limits, path filtering, and concurrent extraction.\n","firecrawl-crawl",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-crawl/SKILL.md",[385],{"path":370,"priority":332},{"basePath":387,"description":388,"displayName":389,"installMethods":390,"rationale":391,"selectedPaths":392,"source":335,"sourceLanguage":18,"type":261},"skills/firecrawl-download","Download an entire website as local files — markdown, screenshots, or multiple formats per page. Use this skill when the user wants to save a site locally, download documentation for offline use, bulk-save pages as files, or says \"download the site\", \"save as local files\", \"offline copy\", \"download all the docs\", or \"save for reference\". Combines site mapping and scraping into organized local directories.\n","firecrawl-download",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-download/SKILL.md",[393],{"path":370,"priority":332},{"basePath":395,"description":396,"displayName":397,"installMethods":398,"rationale":399,"selectedPaths":400,"source":335,"sourceLanguage":18,"type":261},"skills/firecrawl-interact","Control and interact with a live browser session on any scraped page — click buttons, fill forms, navigate flows, and extract data using natural language prompts or code. Use when the user needs to interact with a webpage beyond simple scraping: logging into a site, submitting forms, clicking through pagination, handling infinite scroll, navigating multi-step checkout or wizard flows, or when a regular scrape failed because content is behind JavaScript interaction. Also useful for authenticated scraping via profiles. Triggers on \"interact\", \"click\", \"fill out the form\", \"log in to\", \"sign in\", \"submit\", \"paginated\", \"next page\", \"infinite scroll\", \"interact with the page\", \"navigate to\", \"open a session\", or \"scrape failed\".\n","firecrawl-interact",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-interact/SKILL.md",[401],{"path":370,"priority":332},{"basePath":403,"description":404,"displayName":405,"installMethods":406,"rationale":407,"selectedPaths":408,"source":335,"sourceLanguage":18,"type":261},"skills/firecrawl-map","Discover and list all URLs on a website, with optional search filtering. Use this skill when the user wants to find a specific page on a large site, list all URLs, see the site structure, find where something is on a domain, or says \"map the site\", \"find the URL for\", \"what pages are on\", or \"list all pages\". Essential when the user knows which site but not which exact page.\n","firecrawl-map",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-map/SKILL.md",[409],{"path":370,"priority":332},{"basePath":411,"description":412,"displayName":413,"installMethods":414,"rationale":415,"selectedPaths":416,"source":335,"sourceLanguage":18,"type":261},"skills/firecrawl-parse","Efficiently extract and convert the contents of any local file—such as PDF, DOCX, DOC, ODT, RTF, XLSX, XLS, or HTML—into clean, well-formatted markdown saved to disk. Use this skill whenever the user requests to parse, read, or extract information from a file on their computer, including phrases like “parse this PDF”, “convert this document”, “read this file”, “extract text from”, or when a local file path (not a URL) is provided. This skill offers advanced options like generating AI-powered summaries and answering questions based on the file's content. Prefer this tool over `scrape` when handling local files to deliver precise, structured outputs for downstream tasks.\n","firecrawl-parse",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-parse/SKILL.md",[417],{"path":370,"priority":332},{"basePath":258,"description":10,"displayName":260,"installMethods":419,"rationale":420,"selectedPaths":421,"source":335,"sourceLanguage":18,"type":261},{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-scrape/SKILL.md",[422],{"path":370,"priority":332},{"basePath":424,"description":425,"displayName":426,"installMethods":427,"rationale":428,"selectedPaths":429,"source":335,"sourceLanguage":18,"type":261},"skills/firecrawl-search","Web search with full page content extraction. Use this skill whenever the user asks to search the web, find articles, research a topic, look something up, find recent news, discover sources, or says \"search for\", \"find me\", \"look up\", \"what are people saying about\", or \"find articles about\". Returns real search results with optional full-page markdown — not just snippets. Provides capabilities beyond Claude's built-in WebSearch.\n","firecrawl-search",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-search/SKILL.md",[430],{"path":370,"priority":332},{"basePath":270,"description":432,"displayName":433,"installMethods":434,"license":435,"rationale":436,"selectedPaths":437,"source":335,"sourceLanguage":18,"type":221},"Command-line interface for Firecrawl. Scrape, crawl, and extract data from any website directly from your terminal.","firecrawl-cli",{"npm":433},"ISC","cli ecosystem detected at /",[438,440,441,444,446,448,450,452,454,456,458,460,462,464,466,468,470,472,474,476,478,480,482,484,486,488],{"path":439,"priority":332},"package.json",{"path":334,"priority":332},{"path":442,"priority":443},"src/index.ts","low",{"path":445,"priority":443},"src/commands/agent.ts",{"path":447,"priority":443},"src/commands/browser.ts",{"path":449,"priority":443},"src/commands/config.ts",{"path":451,"priority":443},"src/commands/crawl.ts",{"path":453,"priority":443},"src/commands/create.ts",{"path":455,"priority":443},"src/commands/credit-usage.ts",{"path":457,"priority":443},"src/commands/env.ts",{"path":459,"priority":443},"src/commands/experimental/backends.ts",{"path":461,"priority":443},"src/commands/experimental/index.ts",{"path":463,"priority":443},"src/commands/experimental/shared.ts",{"path":465,"priority":443},"src/commands/init.ts",{"path":467,"priority":443},"src/commands/interact.ts",{"path":469,"priority":443},"src/commands/login.ts",{"path":471,"priority":443},"src/commands/logout.ts",{"path":473,"priority":443},"src/commands/map.ts",{"path":475,"priority":443},"src/commands/parse.ts",{"path":477,"priority":443},"src/commands/scrape.ts",{"path":479,"priority":443},"src/commands/search.ts",{"path":481,"priority":443},"src/commands/setup.ts",{"path":483,"priority":443},"src/commands/skills-install.ts",{"path":485,"priority":443},"src/commands/skills-native.ts",{"path":487,"priority":443},"src/commands/status.ts",{"path":489,"priority":443},"src/commands/version.ts",{"sources":491},[492],"manual",{"npmPackage":433},{"closedIssues90d":245,"description":277,"forks":246,"homepage":495,"openIssues90d":248,"pushedAt":249,"readmeSize":243,"stars":250,"topics":496},"http://docs.firecrawl.dev/cli",[],{"downloads":252},{"classifiedAt":499,"discoverAt":500,"extractAt":501,"githubAt":501,"npmAt":502,"updatedAt":499},1778686940560,1778686934511,1778686936677,1778686938425,[221,224,220,223,219,222,218],{"evaluatedAt":256,"extractAt":302,"updatedAt":256},[],[507,537,563,591,610,629],{"_creationTime":508,"_id":509,"community":510,"display":511,"identity":517,"providers":522,"relations":529,"tags":532,"workflow":533},1778697369634.8499,"k17333g8dacqsmw8kk79h77vmx86n0d1",{"reviewCount":8},{"description":512,"installMethods":513,"name":515,"sourceUrl":516},"AI 代理的浏览器自动化 CLI。当用户需要与网站交互时使用，包括浏览页面、填写表单、点击按钮、截屏、提取数据、测试 Web 应用或自动化任何浏览器任务。触发条件包括请求“打开网站”、“填表”、“点击按钮”、“截屏”、“抓取页面数据”、“测试此 Web 应用”、“登录网站”、“自动化浏览器操作”或任何需要以编程方式进行 Web 交互的任务。",{"claudeCode":514},"shanraisshan/claude-code-best-practice","agent-browser","https://github.com/shanraisshan/claude-code-best-practice",{"basePath":518,"githubOwner":519,"githubRepo":520,"locale":521,"slug":515,"type":261},".claude/skills/agent-browser","shanraisshan","claude-code-best-practice","zh-CN",{"evaluate":523,"extract":528},{"promptVersionExtension":211,"promptVersionScoring":212,"score":524,"tags":525,"targetMarket":225,"tier":290},100,[526,218,527,221,289],"browser-automation","testing",{"commitSha":292},{"repoId":530,"translatedFrom":531},"kd74710g49kxgwbfjxeb7s132d86myxr","k17ej9360g7468qswf3svnrx9186m1ng",[289,526,221,527,218],{"evaluatedAt":534,"extractAt":535,"updatedAt":536},1778697222051,1778697205743,1778697369634,{"_creationTime":538,"_id":539,"community":540,"display":541,"identity":547,"providers":551,"relations":557,"tags":559,"workflow":560},1778696993586.7058,"k174ty9vbfgj7cv3v2szdwjgh986mrsf",{"reviewCount":8},{"description":542,"installMethods":543,"name":545,"sourceUrl":546},"Search ChatGPT and extract the full response + hydration JSON that powers the UI.\nAttaches to a running Chrome instance (port 9222 by default), opens ChatGPT,\nsubmits a query, waits for the streamed response, and returns structured data:\nmessages, product cards, hydration JSON, and API calls.\nUse when asked to \"search chatgpt\", \"ask chatgpt\", \"chatgpt search\",\n\"get chatgpt response\", or \"scrape chatgpt\".\n",{"claudeCode":544},"SeifBenayed/claude-code-sdk","chatgpt-search","https://github.com/SeifBenayed/claude-code-sdk",{"basePath":548,"githubOwner":549,"githubRepo":550,"locale":18,"slug":545,"type":261},".claude/skills/chatgpt-search","SeifBenayed","claude-code-sdk",{"evaluate":552,"extract":556},{"promptVersionExtension":211,"promptVersionScoring":212,"score":524,"tags":553,"targetMarket":225,"tier":290},[554,555,526,220,218],"chatgpt","scraping",{"commitSha":292},{"repoId":558},"kd78s53c1852h5p7c3qem663xs86njab",[526,554,220,555,218],{"evaluatedAt":561,"extractAt":562,"updatedAt":561},1778697026743,1778696993586,{"_creationTime":564,"_id":565,"community":566,"display":567,"identity":573,"providers":577,"relations":584,"tags":587,"workflow":588},1778691104676.009,"k178w7wd1nma48cbwy5hbrnq7s86nyvy",{"reviewCount":8},{"description":568,"installMethods":569,"name":571,"sourceUrl":572},"Extract typed JSON from public website pages using a schema.",{"claudeCode":570},"iterationlayer/skills","website-extraction-api","https://github.com/iterationlayer/skills",{"basePath":574,"githubOwner":575,"githubRepo":576,"locale":18,"slug":571,"type":261},"skills/website-extraction-api","iterationlayer","skills",{"evaluate":578,"extract":583},{"promptVersionExtension":211,"promptVersionScoring":212,"score":524,"tags":579,"targetMarket":225,"tier":290},[218,220,580,581,582,289],"json","schema","api",{"commitSha":292},{"parentExtensionId":585,"repoId":586},"k1721s0xmp59902ybtpakrrffn86n10s","kd76p4g2qmtrkgx99cnab3683d86n4g8",[582,289,220,580,581,218],{"evaluatedAt":589,"extractAt":590,"updatedAt":589},1778694012840,1778691104676,{"_creationTime":592,"_id":593,"community":594,"display":595,"identity":599,"providers":601,"relations":606,"tags":607,"workflow":608},1778691104675.9915,"k172qd89p5z3xybe3h8ncdmns586nd5g",{"reviewCount":8},{"description":596,"installMethods":597,"name":598,"sourceUrl":572},"Extract SKUs, product names, unit prices, availability, and minimum order quantities from a supplier catalog page.",{"claudeCode":570},"extract-supplier-catalog-from-website",{"basePath":600,"githubOwner":575,"githubRepo":576,"locale":18,"slug":598,"type":261},"skills/extract-supplier-catalog-from-website",{"evaluate":602,"extract":605},{"promptVersionExtension":211,"promptVersionScoring":212,"score":524,"tags":603,"targetMarket":225,"tier":290},[218,220,604,582,289],"procurement",{"commitSha":292},{"parentExtensionId":585,"repoId":586},[582,289,220,604,218],{"evaluatedAt":609,"extractAt":590,"updatedAt":609},1778692514878,{"_creationTime":611,"_id":612,"community":613,"display":614,"identity":618,"providers":620,"relations":625,"tags":626,"workflow":627},1778691104675.9885,"k176q61fzr6re9nbpsx4362fw586mvj3",{"reviewCount":8},{"description":615,"installMethods":616,"name":617,"sourceUrl":572},"Extract organization name, registration number, status, registration date, and officers from a public registry page.",{"claudeCode":570},"extract-public-registry-page",{"basePath":619,"githubOwner":575,"githubRepo":576,"locale":18,"slug":617,"type":261},"skills/extract-public-registry-page",{"evaluate":621,"extract":624},{"promptVersionExtension":211,"promptVersionScoring":212,"score":524,"tags":622,"targetMarket":225,"tier":290},[220,218,623,289,582],"compliance",{"commitSha":292},{"parentExtensionId":585,"repoId":586},[582,289,623,220,218],{"evaluatedAt":628,"extractAt":590,"updatedAt":628},1778692237425,{"_creationTime":630,"_id":631,"community":632,"display":633,"identity":639,"providers":643,"relations":650,"tags":653,"workflow":654},1778696691708.3125,"k1704t5mbdhdjkqqyvh15rz98986n4b0",{"reviewCount":8},{"description":634,"installMethods":635,"name":637,"sourceUrl":638},"Extract structured data via stored browser-templates or one-shot DOM queries, with mandatory AIDefence PII + prompt-injection gates before content reaches the model",{"claudeCode":636},"ruvnet/ruflo","browser-extract","https://github.com/ruvnet/ruflo",{"basePath":640,"githubOwner":641,"githubRepo":642,"locale":18,"slug":637,"type":261},"plugins/ruflo-browser/skills/browser-extract","ruvnet","ruflo",{"evaluate":644,"extract":649},{"promptVersionExtension":211,"promptVersionScoring":212,"score":645,"tags":646,"targetMarket":225,"tier":290},99,[218,220,526,647,648],"security","pii-detection",{"commitSha":292},{"parentExtensionId":651,"repoId":652},"k171keecmm9hg25jek1v0vqawd86ntq2","kd7ed28gj8n0y3msk5dzrp05zs86nqtc",[526,220,648,647,218],{"evaluatedAt":655,"extractAt":656,"updatedAt":655},1778699898275,1778696691708]