[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"extension-skill-firecrawl-firecrawl-crawl-de":3,"guides-for-firecrawl-firecrawl-crawl":488,"similar-k173h1mfmqg08wryv3g35k86hn86mmbs-de":489},{"_creationTime":4,"_id":5,"children":6,"community":7,"display":9,"evaluation":15,"identity":241,"isFallback":223,"parentExtension":245,"providers":298,"relations":302,"repo":303,"tags":486,"workflow":487},1778686940775.5713,"k173h1mfmqg08wryv3g35k86hn86mmbs",[],{"reviewCount":8},0,{"description":10,"installMethods":11,"name":13,"sourceUrl":14},"Bulk extract content from an entire website or site section. Use this skill when the user wants to crawl a site, extract all pages from a docs section, bulk-scrape multiple pages following links, or says \"crawl\", \"get all the pages\", \"extract everything under /docs\", \"bulk extract\", or needs content from many pages on the same site. Handles depth limits, path filtering, and concurrent extraction.\n",{"claudeCode":12},"firecrawl/cli","firecrawl-crawl","https://github.com/firecrawl/cli",{"_creationTime":16,"_id":17,"extensionId":5,"locale":18,"result":19,"trustSignals":221,"workflow":239},1778687056905.6401,"kn7ahj9e6hxth5n0vc2ym5echn86m2f0","en",{"checks":20,"evaluatedAt":190,"extensionSummary":191,"features":192,"nonGoals":198,"promptVersionExtension":202,"promptVersionScoring":203,"purpose":204,"rationale":205,"score":206,"summary":207,"tags":208,"targetMarket":214,"tier":215,"useCases":216},[21,26,29,32,36,39,43,46,50,54,58,61,64,67,71,75,78,81,84,88,92,95,99,102,105,108,111,114,117,121,124,129,133,137,140,144,148,152,155,158,161,164,167,170,174,177,180,183,187],{"category":22,"check":23,"severity":24,"summary":25},"Practical Utility","Problem relevance","pass","The description clearly states the problem of bulk content extraction from websites and provides specific use cases like crawling documentation sections or extracting pages following links.",{"category":22,"check":27,"severity":24,"summary":28},"Unique selling proposition","The skill offers value beyond basic web scraping by providing features like depth limits, path filtering, concurrency control, and options for waiting for JS rendering and screenshots, which are advanced capabilities for bulk extraction.",{"category":22,"check":30,"severity":24,"summary":31},"Production readiness","The CLI tool is fully implemented with extensive options for crawling, scraping, searching, and interacting, covering the complete lifecycle of web data extraction and local file saving.",{"category":33,"check":34,"severity":24,"summary":35},"Scope","Single responsibility principle","The extension focuses on web data extraction and manipulation via a CLI, with related commands like scrape, search, map, and crawl all fitting within this domain.",{"category":33,"check":37,"severity":24,"summary":38},"Description quality","The displayed description accurately reflects the skill's capabilities, detailing bulk content extraction, specific use cases, and key features like depth limits and path filtering.",{"category":40,"check":41,"severity":24,"summary":42},"Invocation","Precise Purpose","The purpose is clearly defined as bulk content extraction from websites, with specific triggers and boundaries like crawling documentation sections or extracting pages under '/docs'.",{"category":40,"check":44,"severity":24,"summary":45},"Concise Frontmatter","The SKILL.md frontmatter is concise, self-contained, and accurately summarizes the core capability of bulk content extraction, followed by clear trigger phrases.",{"category":47,"check":48,"severity":24,"summary":49},"Documentation","Concise Body","The SKILL.md body is well-structured with clear headings for different commands and options, effectively delegating detailed procedures to separate files or examples.",{"category":51,"check":52,"severity":24,"summary":53},"Context","Progressive Disclosure","The SKILL.md outlines the main flows and links to separate documentation for detailed explanations, examples, and experimental features, adhering to progressive disclosure.",{"category":51,"check":55,"severity":56,"summary":57},"Forked exploration","not_applicable","The skill does not appear to involve deep exploration or code review that would necessitate context forking.",{"category":22,"check":59,"severity":24,"summary":60},"Usage examples","The documentation provides numerous ready-to-use examples for each command (crawl, scrape, search, agent), demonstrating inputs, invocations, and expected outcomes.",{"category":22,"check":62,"severity":24,"summary":63},"Edge cases","The documentation for commands like `crawl` and `agent` includes relevant options for handling limits, delays, and concurrency, indicating consideration for edge cases and limitations.",{"category":40,"check":65,"severity":56,"summary":66},"Tool Fallback","The skill does not appear to rely on external MCP servers or custom tools that would require fallbacks; it primarily uses its own CLI commands.",{"category":68,"check":69,"severity":24,"summary":70},"Safety","Halt on unexpected state","The commands like `crawl` and `agent` have options for limits, timeouts, and webhook notifications, suggesting mechanisms to halt or manage unexpected states.",{"category":72,"check":73,"severity":24,"summary":74},"Portability","Cross-skill coupling","The skill is self-contained, and while it references other related Firecrawl CLI commands (scrape, map), it does not implicitly rely on other skills being loaded.",{"category":40,"check":76,"severity":24,"summary":77},"Scoped tools","The CLI exposes specific commands like `crawl`, `scrape`, `search`, and `agent`, which are narrow verb-noun specialists, rather than a single generalist tool.",{"category":33,"check":79,"severity":24,"summary":80},"Tool naming","All tool names (`crawl`, `scrape`, `search`, `map`, `agent`, `interact`, `credit-usage`, `config`, `login`, `logout`) are descriptive and easy to understand.",{"category":33,"check":82,"severity":24,"summary":83},"Minimal I/O surface","Commands use specific flags and options for input, and output formats are clearly defined (markdown, JSON, HTML), avoiding excessive or undocumented fields.",{"category":85,"check":86,"severity":24,"summary":87},"License","License usability","The project includes a LICENSE file with the MIT license, which is a permissive open-source license.",{"category":89,"check":90,"severity":24,"summary":91},"Maintenance","Commit recency","The last commit was on May 12, 2026, which is within the last 90 days.",{"category":89,"check":93,"severity":24,"summary":94},"Dependency Management","The project uses npm and has a lockfile, indicating good dependency management practices.",{"category":96,"check":97,"severity":24,"summary":98},"Security","Secret Management","Secrets are handled via API keys and environment variables, with options for per-command keys and self-hosted URLs, and no secrets are echoed in output.",{"category":96,"check":100,"severity":24,"summary":101},"Injection","The CLI handles user input and URLs as data. The 'agent' command specifically uses schemas for structured output, mitigating injection risks.",{"category":96,"check":103,"severity":24,"summary":104},"Transitive Supply-Chain Grenades","The CLI appears to bundle its dependencies and does not perform runtime downloads of code or instructions, mitigating supply-chain risks.",{"category":96,"check":106,"severity":24,"summary":107},"Sandbox Isolation","The CLI operates as a command-line tool and does not appear to modify files outside its designated output paths or project folders.",{"category":96,"check":109,"severity":24,"summary":110},"Sandbox escape primitives","No evidence of detached processes or retry loops around denied tool calls is present in the CLI's structure.",{"category":96,"check":112,"severity":24,"summary":113},"Data Exfiltration","The CLI does not instruct the agent to submit confidential data upstream. Telemetry is opt-out and anonymous.",{"category":96,"check":115,"severity":24,"summary":116},"Hidden Text Tricks","The documentation and code do not contain hidden text tricks or obfuscated instructions.",{"category":118,"check":119,"severity":24,"summary":120},"Hooks","Opaque code execution","The CLI is written in TypeScript and compiled to JavaScript, with no evidence of obfuscated code, base64 payloads, or runtime script fetching.",{"category":72,"check":122,"severity":24,"summary":123},"Structural Assumption","The CLI is designed for command-line use and does not make assumptions about user-specific project organization beyond output directories.",{"category":125,"check":126,"severity":127,"summary":128},"Trust","Issues Attention","warning","In the last 90 days, 4 issues were opened and 1 was closed, indicating a low closure rate and potentially slow maintainer response.",{"category":130,"check":131,"severity":24,"summary":132},"Versioning","Release Management","The CLI has a version number (v1.16.2) indicated in the README and `--version` flag, and a CHANGELOG is present.",{"category":134,"check":135,"severity":24,"summary":136},"Execution","Validation","The CLI uses specific flags and options for input parameters, and the 'agent' command uses JSON schemas for structured output, implying validation.",{"category":96,"check":138,"severity":24,"summary":139},"Unguarded Destructive Operations","The CLI's operations primarily involve data extraction and output to files; destructive operations are not apparent and would likely be guarded.",{"category":141,"check":142,"severity":24,"summary":143},"Errors","Error Handling","The CLI provides clear error messages for authentication failures, invalid arguments, and job statuses, with some suggesting recovery steps.",{"category":145,"check":146,"severity":24,"summary":147},"Code Execution","Logging","The CLI offers options for saving output to files and includes verbose flags like `--progress`, and telemetry is opt-out.",{"category":149,"check":150,"severity":24,"summary":151},"Compliance","GDPR","The tool focuses on public web data. Telemetry is opt-out and anonymous, with no collection of personal data or sensitive command details.",{"category":149,"check":153,"severity":24,"summary":154},"Target market","The tool operates on public web data and is generally applicable globally, with no specific regional limitations detected.",{"category":72,"check":156,"severity":24,"summary":157},"Runtime stability","The CLI is a Node.js application installable via npm, designed to be cross-platform compatible.",{"category":47,"check":159,"severity":24,"summary":160},"README","The README file is comprehensive, detailing installation, authentication, commands, options, examples, and more.",{"category":33,"check":162,"severity":24,"summary":163},"Tool surface size","The CLI exposes a reasonable number of commands (around 10 main commands), fitting within the target range.",{"category":40,"check":165,"severity":24,"summary":166},"Overlapping near-synonym tools","Commands like `crawl` and `map` have distinct purposes, and while `scrape` and `crawl` are related, their scopes are clearly defined.",{"category":47,"check":168,"severity":24,"summary":169},"Phantom features","All advertised features, such as scraping, crawling, and agent capabilities, have corresponding commands and implementations in the CLI.",{"category":171,"check":172,"severity":24,"summary":173},"Install","Installation instruction","The README provides clear installation instructions via npm and `npx`, along with authentication setup and copy-pasteable examples for various commands.",{"category":141,"check":175,"severity":24,"summary":176},"Actionable error messages","Error messages are generally clear, indicating what failed (e.g., authentication, invalid options) and sometimes suggesting remediation (e.g., check API key, use `--help`).",{"category":134,"check":178,"severity":24,"summary":179},"Pinned dependencies","The project includes a `package-lock.json` file, ensuring that third-party dependencies are pinned.",{"category":33,"check":181,"severity":56,"summary":182},"Dry-run preview","The extension is primarily focused on data extraction and output, not state-changing operations that would typically require a dry-run feature.",{"category":184,"check":185,"severity":24,"summary":186},"Protocol","Idempotent retry & timeouts","Options like `--wait`, `--timeout`, `--delay`, and `--max-concurrency` provide control over operation execution and retries, managing external service interactions.",{"category":149,"check":188,"severity":24,"summary":189},"Telemetry opt-in","Telemetry is strictly opt-out via `FIRECRAWL_NO_TELEMETRY=1` environment variable and is described as anonymous and not collecting command data.",1778687056689,"This is a Node.js command-line interface tool for interacting with the Firecrawl API, enabling users to scrape web pages, crawl entire websites, search the web, and run AI agent tasks for data extraction.",[193,194,195,196,197],"Bulk website content extraction","Web crawling with depth and path filtering","Web search with source and category filtering","AI-powered agent for structured data extraction","Interact with scraped pages in a live browser session",[199,200,201],"Replacing a full-fledged browser automation framework for complex UI interactions.","Performing real-time data streaming from dynamic web applications.","Executing arbitrary remote code with elevated privileges.","3.0.0","4.4.0","To provide a robust and flexible command-line interface for bulk content extraction, web crawling, and data discovery from the web.","The primary findings impacting the score are the low issue closure rate and the presence of numerous options that, while functional, could be streamlined. Despite these, the extension is highly practical with extensive documentation and a clear scope.",75,"A powerful and well-documented CLI for web data extraction, crawling, and searching.",[209,210,211,212,213],"web-scraping","cli","data-extraction","automation","web-crawling","global","community",[217,218,219,220],"When you need to crawl a site and extract all pages from a docs section.","When you want to bulk-scrape multiple pages following links on a website.","When performing web searches and needing to scrape results directly.","When automating data extraction tasks from web pages using natural language prompts.",{"codeQuality":222,"collectedAt":224,"documentation":225,"maintenance":228,"popularity":235,"security":237,"testCoverage":238},{"hasLockfile":223},true,1778687034741,{"descriptionLength":226,"readmeSize":227},400,25712,{"closedIssues90d":229,"forks":230,"hasChangelog":231,"openIssues90d":232,"pushedAt":233,"stars":234},1,49,false,4,1778599393000,383,{"npmDownloads":236},51144,{"hasNpmPackage":223,"smitheryVerified":231},{"hasCi":223,"hasTests":223},{"updatedAt":240},1778687056905,{"basePath":242,"githubOwner":243,"githubRepo":210,"locale":18,"slug":13,"type":244},"skills/firecrawl-crawl","firecrawl","skill",{"_creationTime":246,"_id":247,"community":248,"display":249,"identity":252,"parentExtension":255,"providers":286,"relations":294,"tags":295,"workflow":296},1778686940775.5706,"k17axfavjpz72cd3qqzn86shb186ncqt",{"reviewCount":8},{"description":250,"installMethods":251,"name":243,"sourceUrl":14},"Scrape, search, crawl, and map the web with a single command.",{"claudeCode":243},{"basePath":253,"githubOwner":243,"githubRepo":210,"locale":18,"slug":210,"type":254},"","plugin",{"_creationTime":256,"_id":257,"community":258,"display":259,"identity":263,"providers":265,"relations":280,"tags":282,"workflow":283},1778686940775.5703,"k17be4khad1zr773wdg2m4yjhn86mjpr",{"reviewCount":8},{"description":260,"installMethods":261,"name":262,"sourceUrl":14},"CLI and Agent Skill for Firecrawl - Add scrape, search, and browsing capabilities to your AI agents",{"claudeCode":12},"Firecrawl CLI",{"basePath":253,"githubOwner":243,"githubRepo":210,"locale":18,"slug":210,"type":264},"marketplace",{"evaluate":266,"extract":274},{"promptVersionExtension":267,"promptVersionScoring":203,"score":268,"tags":269,"targetMarket":214,"tier":273},"3.1.0",97,[209,210,270,211,271,272,212],"ai-agents","search","crawl","verified",{"commitSha":275,"license":276,"marketplace":277,"plugin":278},"HEAD","MIT",{"name":243,"pluginCount":229},{"mcpCount":8,"provider":279,"skillCount":8},"classify",{"repoId":281},"kd7csd1wb06dg9c1jfy5063f2586ne60",[270,212,210,272,211,271,209],{"evaluatedAt":284,"extractAt":285,"updatedAt":284},1778686962954,1778686940775,{"evaluate":287,"extract":291},{"promptVersionExtension":202,"promptVersionScoring":203,"score":288,"tags":289,"targetMarket":214,"tier":215},78,[209,210,212,211,271,272,290],"agent",{"commitSha":275,"plugin":292},{"mcpCount":8,"provider":279,"skillCount":293},9,{"parentExtensionId":257,"repoId":281},[290,212,210,272,211,271,209],{"evaluatedAt":297,"extractAt":285,"updatedAt":297},1778686985594,{"evaluate":299,"extract":301},{"promptVersionExtension":202,"promptVersionScoring":203,"score":206,"tags":300,"targetMarket":214,"tier":215},[209,210,211,212,213],{"commitSha":275},{"parentExtensionId":247,"repoId":281},{"_creationTime":304,"_id":281,"identity":305,"providers":306,"workflow":481},1778686934511.6108,{"githubOwner":243,"githubRepo":210,"sourceUrl":14},{"classify":307,"discover":473,"extract":476,"github":477,"npm":480},{"commitSha":275,"extensions":308},[309,319,345,354,361,366,374,382,390,398,406,414],{"basePath":253,"displayName":243,"installMethods":310,"rationale":311,"selectedPaths":312,"source":318,"sourceLanguage":18,"type":264},{"claudeCode":12},"marketplace.json at .claude-plugin/marketplace.json",[313,316],{"path":314,"priority":315},".claude-plugin/marketplace.json","mandatory",{"path":317,"priority":315},"README.md","rule",{"basePath":253,"description":250,"displayName":243,"installMethods":320,"rationale":321,"selectedPaths":322,"source":318,"sourceLanguage":18,"type":254},{"claudeCode":243},"plugin manifest at .claude-plugin/plugin.json",[323,325,326,329,331,333,335,337,339,341,343],{"path":324,"priority":315},".claude-plugin/plugin.json",{"path":317,"priority":315},{"path":327,"priority":328},"skills/firecrawl-agent/SKILL.md","medium",{"path":330,"priority":328},"skills/firecrawl-cli/SKILL.md",{"path":332,"priority":328},"skills/firecrawl-crawl/SKILL.md",{"path":334,"priority":328},"skills/firecrawl-download/SKILL.md",{"path":336,"priority":328},"skills/firecrawl-interact/SKILL.md",{"path":338,"priority":328},"skills/firecrawl-map/SKILL.md",{"path":340,"priority":328},"skills/firecrawl-parse/SKILL.md",{"path":342,"priority":328},"skills/firecrawl-scrape/SKILL.md",{"path":344,"priority":328},"skills/firecrawl-search/SKILL.md",{"basePath":346,"description":347,"displayName":348,"installMethods":349,"rationale":350,"selectedPaths":351,"source":318,"sourceLanguage":18,"type":244},"skills/firecrawl-agent","AI-powered autonomous data extraction that navigates complex sites and returns structured JSON. Use this skill when the user wants structured data from websites, needs to extract pricing tiers, product listings, directory entries, or any data as JSON with a schema. Triggers on \"extract structured data\", \"get all the products\", \"pull pricing info\", \"extract as JSON\", or when the user provides a JSON schema for website data. More powerful than simple scraping for multi-page structured extraction.\n","firecrawl-agent",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-agent/SKILL.md",[352],{"path":353,"priority":315},"SKILL.md",{"basePath":355,"description":356,"displayName":243,"installMethods":357,"rationale":358,"selectedPaths":359,"source":318,"sourceLanguage":18,"type":244},"skills/firecrawl-cli","Search, scrape, and interact with the web via the Firecrawl CLI. Use this skill whenever the user wants to search the web, find articles, research a topic, look something up online, scrape a webpage, grab content from a URL, get data from a website, crawl documentation, download a site, or interact with pages that need clicks or logins. Also use when they say \"fetch this page\", \"pull the content from\", \"get the page at https://\", or reference external websites. This provides real-time web search with full page content and interact capabilities — beyond what Claude can do natively with built-in tools. Do NOT trigger for local file operations, git commands, deployments, or code editing tasks.\n",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-cli/SKILL.md",[360],{"path":353,"priority":315},{"basePath":242,"description":10,"displayName":13,"installMethods":362,"rationale":363,"selectedPaths":364,"source":318,"sourceLanguage":18,"type":244},{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-crawl/SKILL.md",[365],{"path":353,"priority":315},{"basePath":367,"description":368,"displayName":369,"installMethods":370,"rationale":371,"selectedPaths":372,"source":318,"sourceLanguage":18,"type":244},"skills/firecrawl-download","Download an entire website as local files — markdown, screenshots, or multiple formats per page. Use this skill when the user wants to save a site locally, download documentation for offline use, bulk-save pages as files, or says \"download the site\", \"save as local files\", \"offline copy\", \"download all the docs\", or \"save for reference\". Combines site mapping and scraping into organized local directories.\n","firecrawl-download",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-download/SKILL.md",[373],{"path":353,"priority":315},{"basePath":375,"description":376,"displayName":377,"installMethods":378,"rationale":379,"selectedPaths":380,"source":318,"sourceLanguage":18,"type":244},"skills/firecrawl-interact","Control and interact with a live browser session on any scraped page — click buttons, fill forms, navigate flows, and extract data using natural language prompts or code. Use when the user needs to interact with a webpage beyond simple scraping: logging into a site, submitting forms, clicking through pagination, handling infinite scroll, navigating multi-step checkout or wizard flows, or when a regular scrape failed because content is behind JavaScript interaction. Also useful for authenticated scraping via profiles. Triggers on \"interact\", \"click\", \"fill out the form\", \"log in to\", \"sign in\", \"submit\", \"paginated\", \"next page\", \"infinite scroll\", \"interact with the page\", \"navigate to\", \"open a session\", or \"scrape failed\".\n","firecrawl-interact",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-interact/SKILL.md",[381],{"path":353,"priority":315},{"basePath":383,"description":384,"displayName":385,"installMethods":386,"rationale":387,"selectedPaths":388,"source":318,"sourceLanguage":18,"type":244},"skills/firecrawl-map","Discover and list all URLs on a website, with optional search filtering. Use this skill when the user wants to find a specific page on a large site, list all URLs, see the site structure, find where something is on a domain, or says \"map the site\", \"find the URL for\", \"what pages are on\", or \"list all pages\". Essential when the user knows which site but not which exact page.\n","firecrawl-map",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-map/SKILL.md",[389],{"path":353,"priority":315},{"basePath":391,"description":392,"displayName":393,"installMethods":394,"rationale":395,"selectedPaths":396,"source":318,"sourceLanguage":18,"type":244},"skills/firecrawl-parse","Efficiently extract and convert the contents of any local file—such as PDF, DOCX, DOC, ODT, RTF, XLSX, XLS, or HTML—into clean, well-formatted markdown saved to disk. Use this skill whenever the user requests to parse, read, or extract information from a file on their computer, including phrases like “parse this PDF”, “convert this document”, “read this file”, “extract text from”, or when a local file path (not a URL) is provided. This skill offers advanced options like generating AI-powered summaries and answering questions based on the file's content. Prefer this tool over `scrape` when handling local files to deliver precise, structured outputs for downstream tasks.\n","firecrawl-parse",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-parse/SKILL.md",[397],{"path":353,"priority":315},{"basePath":399,"description":400,"displayName":401,"installMethods":402,"rationale":403,"selectedPaths":404,"source":318,"sourceLanguage":18,"type":244},"skills/firecrawl-scrape","Extract clean markdown from any URL, including JavaScript-rendered SPAs. Use this skill whenever the user provides a URL and wants its content, says \"scrape\", \"grab\", \"fetch\", \"pull\", \"get the page\", \"extract from this URL\", or \"read this webpage\". Handles JS-rendered pages, multiple concurrent URLs, and returns LLM-optimized markdown. Use this instead of WebFetch for any webpage content extraction.\n","firecrawl-scrape",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-scrape/SKILL.md",[405],{"path":353,"priority":315},{"basePath":407,"description":408,"displayName":409,"installMethods":410,"rationale":411,"selectedPaths":412,"source":318,"sourceLanguage":18,"type":244},"skills/firecrawl-search","Web search with full page content extraction. Use this skill whenever the user asks to search the web, find articles, research a topic, look something up, find recent news, discover sources, or says \"search for\", \"find me\", \"look up\", \"what are people saying about\", or \"find articles about\". Returns real search results with optional full-page markdown — not just snippets. Provides capabilities beyond Claude's built-in WebSearch.\n","firecrawl-search",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-search/SKILL.md",[413],{"path":353,"priority":315},{"basePath":253,"description":415,"displayName":416,"installMethods":417,"license":418,"rationale":419,"selectedPaths":420,"source":318,"sourceLanguage":18,"type":210},"Command-line interface for Firecrawl. Scrape, crawl, and extract data from any website directly from your terminal.","firecrawl-cli",{"npm":416},"ISC","cli ecosystem detected at /",[421,423,424,427,429,431,433,435,437,439,441,443,445,447,449,451,453,455,457,459,461,463,465,467,469,471],{"path":422,"priority":315},"package.json",{"path":317,"priority":315},{"path":425,"priority":426},"src/index.ts","low",{"path":428,"priority":426},"src/commands/agent.ts",{"path":430,"priority":426},"src/commands/browser.ts",{"path":432,"priority":426},"src/commands/config.ts",{"path":434,"priority":426},"src/commands/crawl.ts",{"path":436,"priority":426},"src/commands/create.ts",{"path":438,"priority":426},"src/commands/credit-usage.ts",{"path":440,"priority":426},"src/commands/env.ts",{"path":442,"priority":426},"src/commands/experimental/backends.ts",{"path":444,"priority":426},"src/commands/experimental/index.ts",{"path":446,"priority":426},"src/commands/experimental/shared.ts",{"path":448,"priority":426},"src/commands/init.ts",{"path":450,"priority":426},"src/commands/interact.ts",{"path":452,"priority":426},"src/commands/login.ts",{"path":454,"priority":426},"src/commands/logout.ts",{"path":456,"priority":426},"src/commands/map.ts",{"path":458,"priority":426},"src/commands/parse.ts",{"path":460,"priority":426},"src/commands/scrape.ts",{"path":462,"priority":426},"src/commands/search.ts",{"path":464,"priority":426},"src/commands/setup.ts",{"path":466,"priority":426},"src/commands/skills-install.ts",{"path":468,"priority":426},"src/commands/skills-native.ts",{"path":470,"priority":426},"src/commands/status.ts",{"path":472,"priority":426},"src/commands/version.ts",{"sources":474},[475],"manual",{"npmPackage":416},{"closedIssues90d":229,"description":260,"forks":230,"homepage":478,"openIssues90d":232,"pushedAt":233,"readmeSize":227,"stars":234,"topics":479},"http://docs.firecrawl.dev/cli",[],{"downloads":236},{"classifiedAt":482,"discoverAt":483,"extractAt":484,"githubAt":484,"npmAt":485,"updatedAt":482},1778686940560,1778686934511,1778686936677,1778686938425,[212,210,211,213,209],{"evaluatedAt":240,"extractAt":285,"updatedAt":240},[],[490,520,551,573,599,619],{"_creationTime":491,"_id":492,"community":493,"display":494,"identity":500,"providers":505,"relations":512,"tags":515,"workflow":516},1778697373439.756,"k1735p8vnc44y3ntfepns5m0nx86npbs",{"reviewCount":8},{"description":495,"installMethods":496,"name":498,"sourceUrl":499},"Browser-automatisierungs-CLI für KI-Agenten. Verwenden Sie diese, wenn der Benutzer mit Websites interagieren muss, einschließlich der Navigation auf Seiten, dem Ausfüllen von Formularen, dem Klicken auf Schaltflächen, dem Aufnehmen von Screenshots, dem Extrahieren von Daten, dem Testen von Webanwendungen oder der Automatisierung beliebiger Browseraufgaben. Auslöser sind Anfragen wie \"öffne eine Website\", \"fülle ein Formular aus\", \"klicke auf eine Schaltfläche\", \"mache einen Screenshot\", \"scrappe Daten von einer Seite\", \"teste diese Web-App\", \"melde dich bei einer Seite an\", \"automatisiere Browser-Aktionen\" oder jede Aufgabe, die eine programmatische Webinteraktion erfordert.",{"claudeCode":497},"shanraisshan/claude-code-best-practice","agent-browser","https://github.com/shanraisshan/claude-code-best-practice",{"basePath":501,"githubOwner":502,"githubRepo":503,"locale":504,"slug":498,"type":244},".claude/skills/agent-browser","shanraisshan","claude-code-best-practice","de",{"evaluate":506,"extract":511},{"promptVersionExtension":202,"promptVersionScoring":203,"score":507,"tags":508,"targetMarket":214,"tier":273},100,[509,209,510,210,212],"browser-automation","testing",{"commitSha":275},{"repoId":513,"translatedFrom":514},"kd74710g49kxgwbfjxeb7s132d86myxr","k17ej9360g7468qswf3svnrx9186m1ng",[212,509,210,510,209],{"evaluatedAt":517,"extractAt":518,"updatedAt":519},1778697222051,1778697205743,1778697373439,{"_creationTime":521,"_id":522,"community":523,"display":524,"identity":530,"providers":535,"relations":544,"tags":547,"workflow":548},1778699234184.6157,"k177tdbfgqmwhtaqv771f2ych586nne9",{"reviewCount":8},{"description":525,"installMethods":526,"name":528,"sourceUrl":529},"Worktree-first dev environment manager for issues, PRs, and features with optional tmux sessions",{"claudeCode":527},"Yeachan-Heo/oh-my-claudecode","Project Session Manager","https://github.com/Yeachan-Heo/oh-my-claudecode",{"basePath":531,"githubOwner":532,"githubRepo":533,"locale":18,"slug":534,"type":244},"skills/project-session-manager","Yeachan-Heo","oh-my-claudecode","project-session-manager",{"evaluate":536,"extract":543},{"promptVersionExtension":202,"promptVersionScoring":203,"score":507,"tags":537,"targetMarket":214,"tier":273},[538,539,540,541,212,210,542],"git","development-environment","workflow","tmux","developer-tool",{"commitSha":275,"license":276},{"parentExtensionId":545,"repoId":546},"k17brg5egdw1jbncj1j4wfv3fh86n639","kd74zv63fryf9prygtq7gf4es986n22y",[212,210,542,539,538,541,540],{"evaluatedAt":549,"extractAt":550,"updatedAt":549},1778699613343,1778699234184,{"_creationTime":552,"_id":553,"community":554,"display":555,"identity":558,"providers":561,"relations":569,"tags":570,"workflow":571},1778699234184.611,"k179b6dkc777g1rgyecze04wqn86m6y4",{"reviewCount":8},{"description":556,"installMethods":557,"name":533,"sourceUrl":529},"Process-first advisor routing for Claude, Codex, or Gemini via `omc ask`, with artifact capture and no raw CLI assembly",{"claudeCode":527},{"basePath":559,"githubOwner":532,"githubRepo":533,"locale":18,"slug":560,"type":244},"skills/ask","ask",{"evaluate":562,"extract":568},{"promptVersionExtension":202,"promptVersionScoring":203,"score":507,"tags":563,"targetMarket":214,"tier":273},[210,212,564,565,566,567],"llm","prompting","code-review","artifact-generation",{"commitSha":275,"license":276},{"parentExtensionId":545,"repoId":546},[567,212,210,566,564,565],{"evaluatedAt":572,"extractAt":550,"updatedAt":572},1778699303045,{"_creationTime":574,"_id":575,"community":576,"display":577,"identity":583,"providers":587,"relations":593,"tags":595,"workflow":596},1778697652123.8928,"k171pew5empzzrfghyg9nqrk6n86nqa9",{"reviewCount":8},{"description":578,"installMethods":579,"name":581,"sourceUrl":582},"Use gh for GitHub issues, PR status, CI/logs, comments, reviews, releases, and API queries.",{"claudeCode":580},"steipete/clawdis","github","https://github.com/steipete/clawdis",{"basePath":584,"githubOwner":585,"githubRepo":586,"locale":18,"slug":581,"type":244},"skills/github","steipete","clawdis",{"evaluate":588,"extract":592},{"promptVersionExtension":202,"promptVersionScoring":203,"score":507,"tags":589,"targetMarket":214,"tier":273},[581,210,590,591,212],"api","developer-tools",{"commitSha":275},{"repoId":594},"kd738npxg9yh3xf3vddzy9fyfh86nhng",[590,212,210,591,581],{"evaluatedAt":597,"extractAt":598,"updatedAt":597},1778698569289,1778697652123,{"_creationTime":600,"_id":601,"community":602,"display":603,"identity":607,"providers":609,"relations":615,"tags":616,"workflow":617},1778697652123.8857,"k179k1n5nw0md1x1j3dak31evn86mb5f",{"reviewCount":8},{"description":604,"installMethods":605,"name":606,"sourceUrl":582},"Use when reviewing, reproducing, or proving OpenClaw Telegram behavior with a real Telegram user on Crabbox, including PR review workflows that need an agent-controlled Telegram Desktop recording, TDLib user-driver commands, Convex-leased credentials, WebVNC observation, and motion-trimmed artifacts.",{"claudeCode":580},"telegram-crabbox-e2e-proof",{"basePath":608,"githubOwner":585,"githubRepo":586,"locale":18,"slug":606,"type":244},".agents/skills/telegram-crabbox-e2e-proof",{"evaluate":610,"extract":614},{"promptVersionExtension":202,"promptVersionScoring":203,"score":507,"tags":611,"targetMarket":214,"tier":273},[510,612,212,613,210],"telegram","qa",{"commitSha":275},{"repoId":594},[212,210,613,612,510],{"evaluatedAt":618,"extractAt":598,"updatedAt":618},1778698038113,{"_creationTime":620,"_id":621,"community":622,"display":623,"identity":629,"providers":633,"relations":639,"tags":641,"workflow":642},1778696993586.7058,"k174ty9vbfgj7cv3v2szdwjgh986mrsf",{"reviewCount":8},{"description":624,"installMethods":625,"name":627,"sourceUrl":628},"Search ChatGPT and extract the full response + hydration JSON that powers the UI.\nAttaches to a running Chrome instance (port 9222 by default), opens ChatGPT,\nsubmits a query, waits for the streamed response, and returns structured data:\nmessages, product cards, hydration JSON, and API calls.\nUse when asked to \"search chatgpt\", \"ask chatgpt\", \"chatgpt search\",\n\"get chatgpt response\", or \"scrape chatgpt\".\n",{"claudeCode":626},"SeifBenayed/claude-code-sdk","chatgpt-search","https://github.com/SeifBenayed/claude-code-sdk",{"basePath":630,"githubOwner":631,"githubRepo":632,"locale":18,"slug":627,"type":244},".claude/skills/chatgpt-search","SeifBenayed","claude-code-sdk",{"evaluate":634,"extract":638},{"promptVersionExtension":202,"promptVersionScoring":203,"score":507,"tags":635,"targetMarket":214,"tier":273},[636,637,509,211,209],"chatgpt","scraping",{"commitSha":275},{"repoId":640},"kd78s53c1852h5p7c3qem663xs86njab",[509,636,211,637,209],{"evaluatedAt":643,"extractAt":644,"updatedAt":643},1778697026743,1778696993586]