[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"extension-skill-firecrawl-firecrawl-agent-de":3,"guides-for-firecrawl-firecrawl-agent":492,"similar-k17a0qa1tnktamw8db9gxvtk4986mjbh-de":493},{"_creationTime":4,"_id":5,"children":6,"community":7,"display":9,"evaluation":15,"identity":245,"isFallback":227,"parentExtension":249,"providers":302,"relations":306,"repo":307,"tags":490,"workflow":491},1778686940775.5708,"k17a0qa1tnktamw8db9gxvtk4986mjbh",[],{"reviewCount":8},0,{"description":10,"installMethods":11,"name":13,"sourceUrl":14},"AI-powered autonomous data extraction that navigates complex sites and returns structured JSON. Use this skill when the user wants structured data from websites, needs to extract pricing tiers, product listings, directory entries, or any data as JSON with a schema. Triggers on \"extract structured data\", \"get all the products\", \"pull pricing info\", \"extract as JSON\", or when the user provides a JSON schema for website data. More powerful than simple scraping for multi-page structured extraction.\n",{"claudeCode":12},"firecrawl/cli","firecrawl-agent","https://github.com/firecrawl/cli",{"_creationTime":16,"_id":17,"extensionId":5,"locale":18,"result":19,"trustSignals":225,"workflow":243},1778687006207.021,"kn78py90tb3gw2z640rqd4x5y986mmj9","en",{"checks":20,"evaluatedAt":193,"extensionSummary":194,"features":195,"nonGoals":201,"promptVersionExtension":205,"promptVersionScoring":206,"purpose":207,"rationale":208,"score":209,"summary":210,"tags":211,"targetMarket":218,"tier":219,"useCases":220},[21,26,29,32,36,39,44,48,51,54,58,62,65,69,72,75,78,81,84,87,91,95,100,104,108,111,115,118,122,125,128,131,134,137,140,144,148,151,154,158,161,164,167,170,174,177,180,183,186,190],{"category":22,"check":23,"severity":24,"summary":25},"Practical Utility","Problem relevance","pass","The description clearly identifies the problem of needing structured data from complex websites and positions the skill as a solution for extracting specific data types like pricing tiers and product listings.",{"category":22,"check":27,"severity":24,"summary":28},"Unique selling proposition","The skill offers significant value beyond a simple prompt by autonomously navigating complex sites, handling multi-page extractions, and returning structured JSON, which is a distinct capability.",{"category":22,"check":30,"severity":24,"summary":31},"Production readiness","The skill provides autonomous data extraction for complex sites, returning structured JSON, and the CLI includes options for waiting for completion and specifying output, indicating readiness for real workflows.",{"category":33,"check":34,"severity":24,"summary":35},"Scope","Single responsibility principle","The skill focuses specifically on AI-powered autonomous data extraction from websites, returning structured JSON, and does not appear to encompass unrelated domains.",{"category":33,"check":37,"severity":24,"summary":38},"Description quality","The displayed description accurately reflects the skill's capability for AI-powered autonomous data extraction, navigation of complex sites, and structured JSON output, including relevant triggers and use cases.",{"category":40,"check":41,"severity":42,"summary":43},"Invocation","Scoped tools","not_applicable","This is a skill, not a tool-based extension, so the concept of scoped tools does not apply.",{"category":45,"check":46,"severity":24,"summary":47},"Documentation","Configuration & parameter reference","The SKILL.md and README.md provide a comprehensive list of options and parameters for the various commands, including descriptions and default values where applicable.",{"category":33,"check":49,"severity":24,"summary":50},"Tool naming","The command names within the CLI (e.g., `scrape`, `search`, `agent`) are descriptive and clearly indicate their function.",{"category":33,"check":52,"severity":24,"summary":53},"Minimal I/O surface","The CLI commands accept structured inputs via flags and schema definitions, and outputs are focused on the requested data, avoiding excessive diagnostic information.",{"category":55,"check":56,"severity":24,"summary":57},"License","License usability","The project's license is available via a LICENSE file and is a permissive MIT license, suitable for broad use.",{"category":59,"check":60,"severity":24,"summary":61},"Maintenance","Commit recency","The last commit was on May 12, 2026, indicating recent maintenance.",{"category":59,"check":63,"severity":24,"summary":64},"Dependency Management","The project uses npm and has a lockfile, indicating proper dependency management.",{"category":66,"check":67,"severity":24,"summary":68},"Security","Secret Management","The CLI handles API keys via environment variables or command-line arguments and does not appear to hardcode secrets or expose them insecurely.",{"category":66,"check":70,"severity":24,"summary":71},"Injection","The CLI commands are designed to process data as input and do not execute arbitrary instructions from external sources.",{"category":66,"check":73,"severity":24,"summary":74},"Transitive Supply-Chain Grenades","The CLI appears to bundle dependencies and does not execute remote scripts or fetch content as instructions at runtime.",{"category":66,"check":76,"severity":24,"summary":77},"Sandbox Isolation","The CLI operates as a command-line tool and does not appear to modify files outside of its intended output directories or project scopes.",{"category":66,"check":79,"severity":24,"summary":80},"Sandbox escape primitives","No detached process spawns or deny-retry loops were observed in the provided script snippets or documentation.",{"category":66,"check":82,"severity":24,"summary":83},"Data Exfiltration","The CLI's telemetry is opt-in and explicitly states that command data, URLs, or file contents are not collected. Outbound calls are primarily to the Firecrawl API.",{"category":66,"check":85,"severity":24,"summary":86},"Hidden Text Tricks","The bundled content and documentation do not contain hidden steering tricks or malicious invisible characters.",{"category":88,"check":89,"severity":24,"summary":90},"Hooks","Opaque code execution","The CLI commands are standard shell commands and do not involve obfuscated code execution or runtime script fetching.",{"category":92,"check":93,"severity":24,"summary":94},"Portability","Structural Assumption","The CLI commands operate on provided URLs or files and do not make assumptions about specific user project directory structures.",{"category":96,"check":97,"severity":98,"summary":99},"Trust","Issues Attention","warning","In the last 90 days, 4 issues were opened and 1 was closed, indicating slow issue resolution.",{"category":101,"check":102,"severity":24,"summary":103},"Versioning","Release Management","The CLI has a version number (v1.16.2) and a CHANGELOG is referenced, indicating clear release management.",{"category":105,"check":106,"severity":24,"summary":107},"Execution","Validation","The CLI's options and parameters appear to be validated, and schema-based extraction is supported, suggesting robust input handling.",{"category":66,"check":109,"severity":24,"summary":110},"Unguarded Destructive Operations","The primary operations are data extraction and scraping, which are not inherently destructive. Commands that modify local files use output paths and do not perform irreversible actions without user specification.",{"category":112,"check":113,"severity":24,"summary":114},"Code Execution","Error Handling","The CLI provides clear error messages for authentication failures, invalid commands, and job execution issues, with some guidance on remediation.",{"category":112,"check":116,"severity":24,"summary":117},"Logging","The CLI outputs results to stdout or specified files, and provides status updates and error messages, allowing for review of executed actions.",{"category":119,"check":120,"severity":24,"summary":121},"Compliance","GDPR","The CLI focuses on scraping public web data and does not explicitly operate on personal data without sanitization. Telemetry is opt-in.",{"category":119,"check":123,"severity":24,"summary":124},"Target market","The CLI is designed for general web scraping and data extraction and does not exhibit regional or jurisdictional restrictions, hence it is global.",{"category":92,"check":126,"severity":24,"summary":127},"Runtime stability","The CLI is a Node.js application installable via npm and is designed to run on standard environments, with no apparent OS-specific dependencies.",{"category":45,"check":129,"severity":24,"summary":130},"README","The README file is comprehensive, detailing installation, authentication, commands, options, and usage examples.",{"category":33,"check":132,"severity":24,"summary":133},"Tool surface size","The CLI exposes a focused set of commands (scrape, search, map, crawl, agent, interact, config, etc.), well within the recommended range.",{"category":40,"check":135,"severity":24,"summary":136},"Overlapping near-synonym tools","The commands are distinct (e.g., `scrape` vs. `crawl` vs. `agent`) and do not appear to have significant overlap in their core functionality.",{"category":45,"check":138,"severity":24,"summary":139},"Phantom features","All advertised features, such as scraping, searching, and agent-based extraction, have corresponding CLI commands and documentation.",{"category":141,"check":142,"severity":24,"summary":143},"Install","Installation instruction","The README provides clear installation instructions via npm and includes authentication setup steps and copy-pasteable command examples.",{"category":145,"check":146,"severity":24,"summary":147},"Errors","Actionable error messages","The CLI provides informative error messages that generally indicate the operation that failed and often suggest next steps or point to documentation.",{"category":105,"check":149,"severity":24,"summary":150},"Pinned dependencies","The project utilizes npm with a lockfile, ensuring that third-party dependencies are pinned.",{"category":33,"check":152,"severity":42,"summary":153},"Dry-run preview","The primary functions of the CLI are data extraction and retrieval, which are not state-changing operations requiring a dry-run mode.",{"category":155,"check":156,"severity":24,"summary":157},"Protocol","Idempotent retry & timeouts","While not explicitly detailed for all operations, the CLI supports waiting for jobs and has timeout options for agent tasks, suggesting some consideration for operation completion. The core operations are retrieval-based.",{"category":119,"check":159,"severity":24,"summary":160},"Telemetry opt-in","Telemetry is strictly opt-in via an environment variable (`FIRECRAWL_NO_TELEMETRY=1`) and its scope is documented.",{"category":40,"check":162,"severity":24,"summary":163},"Precise Purpose","The skill's purpose is clearly defined as AI-powered autonomous data extraction with a focus on complex sites and structured JSON output, with specific triggers and boundaries outlined.",{"category":40,"check":165,"severity":24,"summary":166},"Concise Frontmatter","The frontmatter in SKILL.md is concise, self-contained, and clearly summarizes the core capability and trigger phrases.",{"category":45,"check":168,"severity":24,"summary":169},"Concise Body","The SKILL.md file is reasonably concise, delegating more detailed information to separate markdown files and the README.",{"category":171,"check":172,"severity":24,"summary":173},"Context","Progressive Disclosure","The SKILL.md directs users to the README for installation and broader usage, and the README references the full documentation for more details, demonstrating progressive disclosure.",{"category":171,"check":175,"severity":42,"summary":176},"Forked exploration","This skill is a CLI tool focused on data extraction and does not involve deep exploration or code review requiring a forked context.",{"category":22,"check":178,"severity":24,"summary":179},"Usage examples","The README and SKILL.md provide numerous clear, ready-to-use examples for all major commands and their options.",{"category":22,"check":181,"severity":24,"summary":182},"Edge cases","The documentation addresses limitations such as agent task duration, credit limits, and suggests alternatives for simpler tasks, indirectly covering some failure modes.",{"category":112,"check":184,"severity":42,"summary":185},"Tool Fallback","The skill does not explicitly rely on external MCP servers or tools that would require a fallback mechanism.",{"category":187,"check":188,"severity":24,"summary":189},"Safety","Halt on unexpected state","The CLI's operations are primarily focused on data retrieval and do not involve destructive actions that would require complex state management or pre-state checks beyond basic validation.",{"category":92,"check":191,"severity":42,"summary":192},"Cross-skill coupling","As a CLI tool, it does not implicitly rely on other skills being loaded in the same session.",1778687006089,"This is a command-line interface (CLI) tool for Firecrawl that enables AI-powered autonomous data extraction from complex websites. It can navigate multi-page sites, extract structured data into JSON format based on prompts or provided schemas, and offers various commands for searching, scraping, mapping URLs, crawling websites, and interacting with live browser sessions.",[196,197,198,199,200],"AI-powered autonomous data extraction","Navigates complex and multi-page sites","Returns structured JSON output","Supports custom JSON schemas for extraction","Provides CLI commands for search, scrape, crawl, and interaction",[202,203,204],"Simple single-page scraping (prefers `firecrawl-scrape` for cost/speed)","Real-time data monitoring or alerting","Replacing dedicated database or data warehousing solutions","3.0.0","4.4.0","To autonomously extract structured data from complex websites and return it as JSON, especially when manual scraping is inefficient.","The extension is well-documented, production-ready, and addresses a clear user problem with a unique selling proposition. The primary finding impacting the score is the slow issue resolution rate.",95,"A highly capable AI-powered CLI for autonomous web data extraction and structured JSON generation.",[212,213,214,215,216,217],"data-extraction","web-scraping","json","cli","ai","automation","global","community",[221,222,223,224],"Extracting pricing tiers from e-commerce sites","Pulling product listings with structured details","Gathering directory entries with specific fields","When structured data is needed from complex websites that require navigation",{"codeQuality":226,"collectedAt":228,"documentation":229,"maintenance":232,"popularity":239,"security":241,"testCoverage":242},{"hasLockfile":227},true,1778686985845,{"descriptionLength":230,"readmeSize":231},500,25712,{"closedIssues90d":233,"forks":234,"hasChangelog":235,"openIssues90d":236,"pushedAt":237,"stars":238},1,49,false,4,1778599393000,383,{"npmDownloads":240},51144,{"hasNpmPackage":227,"smitheryVerified":235},{"hasCi":227,"hasTests":227},{"updatedAt":244},1778687006207,{"basePath":246,"githubOwner":247,"githubRepo":215,"locale":18,"slug":13,"type":248},"skills/firecrawl-agent","firecrawl","skill",{"_creationTime":250,"_id":251,"community":252,"display":253,"identity":256,"parentExtension":259,"providers":290,"relations":298,"tags":299,"workflow":300},1778686940775.5706,"k17axfavjpz72cd3qqzn86shb186ncqt",{"reviewCount":8},{"description":254,"installMethods":255,"name":247,"sourceUrl":14},"Scrape, search, crawl, and map the web with a single command.",{"claudeCode":247},{"basePath":257,"githubOwner":247,"githubRepo":215,"locale":18,"slug":215,"type":258},"","plugin",{"_creationTime":260,"_id":261,"community":262,"display":263,"identity":267,"providers":269,"relations":284,"tags":286,"workflow":287},1778686940775.5703,"k17be4khad1zr773wdg2m4yjhn86mjpr",{"reviewCount":8},{"description":264,"installMethods":265,"name":266,"sourceUrl":14},"CLI and Agent Skill for Firecrawl - Add scrape, search, and browsing capabilities to your AI agents",{"claudeCode":12},"Firecrawl CLI",{"basePath":257,"githubOwner":247,"githubRepo":215,"locale":18,"slug":215,"type":268},"marketplace",{"evaluate":270,"extract":278},{"promptVersionExtension":271,"promptVersionScoring":206,"score":272,"tags":273,"targetMarket":218,"tier":277},"3.1.0",97,[213,215,274,212,275,276,217],"ai-agents","search","crawl","verified",{"commitSha":279,"license":280,"marketplace":281,"plugin":282},"HEAD","MIT",{"name":247,"pluginCount":233},{"mcpCount":8,"provider":283,"skillCount":8},"classify",{"repoId":285},"kd7csd1wb06dg9c1jfy5063f2586ne60",[274,217,215,276,212,275,213],{"evaluatedAt":288,"extractAt":289,"updatedAt":288},1778686962954,1778686940775,{"evaluate":291,"extract":295},{"promptVersionExtension":205,"promptVersionScoring":206,"score":292,"tags":293,"targetMarket":218,"tier":219},78,[213,215,217,212,275,276,294],"agent",{"commitSha":279,"plugin":296},{"mcpCount":8,"provider":283,"skillCount":297},9,{"parentExtensionId":261,"repoId":285},[294,217,215,276,212,275,213],{"evaluatedAt":301,"extractAt":289,"updatedAt":301},1778686985594,{"evaluate":303,"extract":305},{"promptVersionExtension":205,"promptVersionScoring":206,"score":209,"tags":304,"targetMarket":218,"tier":219},[212,213,214,215,216,217],{"commitSha":279},{"parentExtensionId":251,"repoId":285},{"_creationTime":308,"_id":285,"identity":309,"providers":310,"workflow":485},1778686934511.6108,{"githubOwner":247,"githubRepo":215,"sourceUrl":14},{"classify":311,"discover":477,"extract":480,"github":481,"npm":484},{"commitSha":279,"extensions":312},[313,323,349,355,362,370,378,386,394,402,410,418],{"basePath":257,"displayName":247,"installMethods":314,"rationale":315,"selectedPaths":316,"source":322,"sourceLanguage":18,"type":268},{"claudeCode":12},"marketplace.json at .claude-plugin/marketplace.json",[317,320],{"path":318,"priority":319},".claude-plugin/marketplace.json","mandatory",{"path":321,"priority":319},"README.md","rule",{"basePath":257,"description":254,"displayName":247,"installMethods":324,"rationale":325,"selectedPaths":326,"source":322,"sourceLanguage":18,"type":258},{"claudeCode":247},"plugin manifest at .claude-plugin/plugin.json",[327,329,330,333,335,337,339,341,343,345,347],{"path":328,"priority":319},".claude-plugin/plugin.json",{"path":321,"priority":319},{"path":331,"priority":332},"skills/firecrawl-agent/SKILL.md","medium",{"path":334,"priority":332},"skills/firecrawl-cli/SKILL.md",{"path":336,"priority":332},"skills/firecrawl-crawl/SKILL.md",{"path":338,"priority":332},"skills/firecrawl-download/SKILL.md",{"path":340,"priority":332},"skills/firecrawl-interact/SKILL.md",{"path":342,"priority":332},"skills/firecrawl-map/SKILL.md",{"path":344,"priority":332},"skills/firecrawl-parse/SKILL.md",{"path":346,"priority":332},"skills/firecrawl-scrape/SKILL.md",{"path":348,"priority":332},"skills/firecrawl-search/SKILL.md",{"basePath":246,"description":10,"displayName":13,"installMethods":350,"rationale":351,"selectedPaths":352,"source":322,"sourceLanguage":18,"type":248},{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-agent/SKILL.md",[353],{"path":354,"priority":319},"SKILL.md",{"basePath":356,"description":357,"displayName":247,"installMethods":358,"rationale":359,"selectedPaths":360,"source":322,"sourceLanguage":18,"type":248},"skills/firecrawl-cli","Search, scrape, and interact with the web via the Firecrawl CLI. Use this skill whenever the user wants to search the web, find articles, research a topic, look something up online, scrape a webpage, grab content from a URL, get data from a website, crawl documentation, download a site, or interact with pages that need clicks or logins. Also use when they say \"fetch this page\", \"pull the content from\", \"get the page at https://\", or reference external websites. This provides real-time web search with full page content and interact capabilities — beyond what Claude can do natively with built-in tools. Do NOT trigger for local file operations, git commands, deployments, or code editing tasks.\n",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-cli/SKILL.md",[361],{"path":354,"priority":319},{"basePath":363,"description":364,"displayName":365,"installMethods":366,"rationale":367,"selectedPaths":368,"source":322,"sourceLanguage":18,"type":248},"skills/firecrawl-crawl","Bulk extract content from an entire website or site section. Use this skill when the user wants to crawl a site, extract all pages from a docs section, bulk-scrape multiple pages following links, or says \"crawl\", \"get all the pages\", \"extract everything under /docs\", \"bulk extract\", or needs content from many pages on the same site. Handles depth limits, path filtering, and concurrent extraction.\n","firecrawl-crawl",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-crawl/SKILL.md",[369],{"path":354,"priority":319},{"basePath":371,"description":372,"displayName":373,"installMethods":374,"rationale":375,"selectedPaths":376,"source":322,"sourceLanguage":18,"type":248},"skills/firecrawl-download","Download an entire website as local files — markdown, screenshots, or multiple formats per page. Use this skill when the user wants to save a site locally, download documentation for offline use, bulk-save pages as files, or says \"download the site\", \"save as local files\", \"offline copy\", \"download all the docs\", or \"save for reference\". Combines site mapping and scraping into organized local directories.\n","firecrawl-download",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-download/SKILL.md",[377],{"path":354,"priority":319},{"basePath":379,"description":380,"displayName":381,"installMethods":382,"rationale":383,"selectedPaths":384,"source":322,"sourceLanguage":18,"type":248},"skills/firecrawl-interact","Control and interact with a live browser session on any scraped page — click buttons, fill forms, navigate flows, and extract data using natural language prompts or code. Use when the user needs to interact with a webpage beyond simple scraping: logging into a site, submitting forms, clicking through pagination, handling infinite scroll, navigating multi-step checkout or wizard flows, or when a regular scrape failed because content is behind JavaScript interaction. Also useful for authenticated scraping via profiles. Triggers on \"interact\", \"click\", \"fill out the form\", \"log in to\", \"sign in\", \"submit\", \"paginated\", \"next page\", \"infinite scroll\", \"interact with the page\", \"navigate to\", \"open a session\", or \"scrape failed\".\n","firecrawl-interact",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-interact/SKILL.md",[385],{"path":354,"priority":319},{"basePath":387,"description":388,"displayName":389,"installMethods":390,"rationale":391,"selectedPaths":392,"source":322,"sourceLanguage":18,"type":248},"skills/firecrawl-map","Discover and list all URLs on a website, with optional search filtering. Use this skill when the user wants to find a specific page on a large site, list all URLs, see the site structure, find where something is on a domain, or says \"map the site\", \"find the URL for\", \"what pages are on\", or \"list all pages\". Essential when the user knows which site but not which exact page.\n","firecrawl-map",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-map/SKILL.md",[393],{"path":354,"priority":319},{"basePath":395,"description":396,"displayName":397,"installMethods":398,"rationale":399,"selectedPaths":400,"source":322,"sourceLanguage":18,"type":248},"skills/firecrawl-parse","Efficiently extract and convert the contents of any local file—such as PDF, DOCX, DOC, ODT, RTF, XLSX, XLS, or HTML—into clean, well-formatted markdown saved to disk. Use this skill whenever the user requests to parse, read, or extract information from a file on their computer, including phrases like “parse this PDF”, “convert this document”, “read this file”, “extract text from”, or when a local file path (not a URL) is provided. This skill offers advanced options like generating AI-powered summaries and answering questions based on the file's content. Prefer this tool over `scrape` when handling local files to deliver precise, structured outputs for downstream tasks.\n","firecrawl-parse",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-parse/SKILL.md",[401],{"path":354,"priority":319},{"basePath":403,"description":404,"displayName":405,"installMethods":406,"rationale":407,"selectedPaths":408,"source":322,"sourceLanguage":18,"type":248},"skills/firecrawl-scrape","Extract clean markdown from any URL, including JavaScript-rendered SPAs. Use this skill whenever the user provides a URL and wants its content, says \"scrape\", \"grab\", \"fetch\", \"pull\", \"get the page\", \"extract from this URL\", or \"read this webpage\". Handles JS-rendered pages, multiple concurrent URLs, and returns LLM-optimized markdown. Use this instead of WebFetch for any webpage content extraction.\n","firecrawl-scrape",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-scrape/SKILL.md",[409],{"path":354,"priority":319},{"basePath":411,"description":412,"displayName":413,"installMethods":414,"rationale":415,"selectedPaths":416,"source":322,"sourceLanguage":18,"type":248},"skills/firecrawl-search","Web search with full page content extraction. Use this skill whenever the user asks to search the web, find articles, research a topic, look something up, find recent news, discover sources, or says \"search for\", \"find me\", \"look up\", \"what are people saying about\", or \"find articles about\". Returns real search results with optional full-page markdown — not just snippets. Provides capabilities beyond Claude's built-in WebSearch.\n","firecrawl-search",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-search/SKILL.md",[417],{"path":354,"priority":319},{"basePath":257,"description":419,"displayName":420,"installMethods":421,"license":422,"rationale":423,"selectedPaths":424,"source":322,"sourceLanguage":18,"type":215},"Command-line interface for Firecrawl. Scrape, crawl, and extract data from any website directly from your terminal.","firecrawl-cli",{"npm":420},"ISC","cli ecosystem detected at /",[425,427,428,431,433,435,437,439,441,443,445,447,449,451,453,455,457,459,461,463,465,467,469,471,473,475],{"path":426,"priority":319},"package.json",{"path":321,"priority":319},{"path":429,"priority":430},"src/index.ts","low",{"path":432,"priority":430},"src/commands/agent.ts",{"path":434,"priority":430},"src/commands/browser.ts",{"path":436,"priority":430},"src/commands/config.ts",{"path":438,"priority":430},"src/commands/crawl.ts",{"path":440,"priority":430},"src/commands/create.ts",{"path":442,"priority":430},"src/commands/credit-usage.ts",{"path":444,"priority":430},"src/commands/env.ts",{"path":446,"priority":430},"src/commands/experimental/backends.ts",{"path":448,"priority":430},"src/commands/experimental/index.ts",{"path":450,"priority":430},"src/commands/experimental/shared.ts",{"path":452,"priority":430},"src/commands/init.ts",{"path":454,"priority":430},"src/commands/interact.ts",{"path":456,"priority":430},"src/commands/login.ts",{"path":458,"priority":430},"src/commands/logout.ts",{"path":460,"priority":430},"src/commands/map.ts",{"path":462,"priority":430},"src/commands/parse.ts",{"path":464,"priority":430},"src/commands/scrape.ts",{"path":466,"priority":430},"src/commands/search.ts",{"path":468,"priority":430},"src/commands/setup.ts",{"path":470,"priority":430},"src/commands/skills-install.ts",{"path":472,"priority":430},"src/commands/skills-native.ts",{"path":474,"priority":430},"src/commands/status.ts",{"path":476,"priority":430},"src/commands/version.ts",{"sources":478},[479],"manual",{"npmPackage":420},{"closedIssues90d":233,"description":264,"forks":234,"homepage":482,"openIssues90d":236,"pushedAt":237,"readmeSize":231,"stars":238,"topics":483},"http://docs.firecrawl.dev/cli",[],{"downloads":240},{"classifiedAt":486,"discoverAt":487,"extractAt":488,"githubAt":488,"npmAt":489,"updatedAt":486},1778686940560,1778686934511,1778686936677,1778686938425,[216,217,215,212,214,213],{"evaluatedAt":244,"extractAt":289,"updatedAt":244},[],[494,524,551,581,612,634],{"_creationTime":495,"_id":496,"community":497,"display":498,"identity":504,"providers":509,"relations":516,"tags":519,"workflow":520},1778697373439.756,"k1735p8vnc44y3ntfepns5m0nx86npbs",{"reviewCount":8},{"description":499,"installMethods":500,"name":502,"sourceUrl":503},"Browser-automatisierungs-CLI für KI-Agenten. Verwenden Sie diese, wenn der Benutzer mit Websites interagieren muss, einschließlich der Navigation auf Seiten, dem Ausfüllen von Formularen, dem Klicken auf Schaltflächen, dem Aufnehmen von Screenshots, dem Extrahieren von Daten, dem Testen von Webanwendungen oder der Automatisierung beliebiger Browseraufgaben. Auslöser sind Anfragen wie \"öffne eine Website\", \"fülle ein Formular aus\", \"klicke auf eine Schaltfläche\", \"mache einen Screenshot\", \"scrappe Daten von einer Seite\", \"teste diese Web-App\", \"melde dich bei einer Seite an\", \"automatisiere Browser-Aktionen\" oder jede Aufgabe, die eine programmatische Webinteraktion erfordert.",{"claudeCode":501},"shanraisshan/claude-code-best-practice","agent-browser","https://github.com/shanraisshan/claude-code-best-practice",{"basePath":505,"githubOwner":506,"githubRepo":507,"locale":508,"slug":502,"type":248},".claude/skills/agent-browser","shanraisshan","claude-code-best-practice","de",{"evaluate":510,"extract":515},{"promptVersionExtension":205,"promptVersionScoring":206,"score":511,"tags":512,"targetMarket":218,"tier":277},100,[513,213,514,215,217],"browser-automation","testing",{"commitSha":279},{"repoId":517,"translatedFrom":518},"kd74710g49kxgwbfjxeb7s132d86myxr","k17ej9360g7468qswf3svnrx9186m1ng",[217,513,215,514,213],{"evaluatedAt":521,"extractAt":522,"updatedAt":523},1778697222051,1778697205743,1778697373439,{"_creationTime":525,"_id":526,"community":527,"display":528,"identity":534,"providers":538,"relations":544,"tags":547,"workflow":548},1778691104676.009,"k178w7wd1nma48cbwy5hbrnq7s86nyvy",{"reviewCount":8},{"description":529,"installMethods":530,"name":532,"sourceUrl":533},"Extract typed JSON from public website pages using a schema.",{"claudeCode":531},"iterationlayer/skills","website-extraction-api","https://github.com/iterationlayer/skills",{"basePath":535,"githubOwner":536,"githubRepo":537,"locale":18,"slug":532,"type":248},"skills/website-extraction-api","iterationlayer","skills",{"evaluate":539,"extract":543},{"promptVersionExtension":205,"promptVersionScoring":206,"score":511,"tags":540,"targetMarket":218,"tier":277},[213,212,214,541,542,217],"schema","api",{"commitSha":279},{"parentExtensionId":545,"repoId":546},"k1721s0xmp59902ybtpakrrffn86n10s","kd76p4g2qmtrkgx99cnab3683d86n4g8",[542,217,212,214,541,213],{"evaluatedAt":549,"extractAt":550,"updatedAt":549},1778694012840,1778691104676,{"_creationTime":552,"_id":553,"community":554,"display":555,"identity":561,"providers":566,"relations":574,"tags":577,"workflow":578},1778696691708.3274,"k170az7r02e9e2v47mpy80kx6n86nff3",{"reviewCount":8},{"description":556,"installMethods":557,"name":559,"sourceUrl":560},"Detect current market regime using npx neural-trader — bull/bear/ranging/volatile classification with recommended strategy",{"claudeCode":558},"ruvnet/ruflo","Trader Regime","https://github.com/ruvnet/ruflo",{"basePath":562,"githubOwner":563,"githubRepo":564,"locale":18,"slug":565,"type":248},"plugins/ruflo-neural-trader/skills/trader-regime","ruvnet","ruflo","trader-regime",{"evaluate":567,"extract":573},{"promptVersionExtension":205,"promptVersionScoring":206,"score":511,"tags":568,"targetMarket":218,"tier":277},[569,570,571,216,572,215],"finance","trading","market-analysis","typescript",{"commitSha":279,"license":280},{"parentExtensionId":575,"repoId":576},"k17drge8h1fgzchr0p4jaeg33n86mwmy","kd7ed28gj8n0y3msk5dzrp05zs86nqtc",[216,215,569,571,570,572],{"evaluatedAt":579,"extractAt":580,"updatedAt":579},1778701108877,1778696691708,{"_creationTime":582,"_id":583,"community":584,"display":585,"identity":591,"providers":596,"relations":605,"tags":608,"workflow":609},1778699234184.6157,"k177tdbfgqmwhtaqv771f2ych586nne9",{"reviewCount":8},{"description":586,"installMethods":587,"name":589,"sourceUrl":590},"Worktree-first dev environment manager for issues, PRs, and features with optional tmux sessions",{"claudeCode":588},"Yeachan-Heo/oh-my-claudecode","Project Session Manager","https://github.com/Yeachan-Heo/oh-my-claudecode",{"basePath":592,"githubOwner":593,"githubRepo":594,"locale":18,"slug":595,"type":248},"skills/project-session-manager","Yeachan-Heo","oh-my-claudecode","project-session-manager",{"evaluate":597,"extract":604},{"promptVersionExtension":205,"promptVersionScoring":206,"score":511,"tags":598,"targetMarket":218,"tier":277},[599,600,601,602,217,215,603],"git","development-environment","workflow","tmux","developer-tool",{"commitSha":279,"license":280},{"parentExtensionId":606,"repoId":607},"k17brg5egdw1jbncj1j4wfv3fh86n639","kd74zv63fryf9prygtq7gf4es986n22y",[217,215,603,600,599,602,601],{"evaluatedAt":610,"extractAt":611,"updatedAt":610},1778699613343,1778699234184,{"_creationTime":613,"_id":614,"community":615,"display":616,"identity":619,"providers":622,"relations":630,"tags":631,"workflow":632},1778699234184.611,"k179b6dkc777g1rgyecze04wqn86m6y4",{"reviewCount":8},{"description":617,"installMethods":618,"name":594,"sourceUrl":590},"Process-first advisor routing for Claude, Codex, or Gemini via `omc ask`, with artifact capture and no raw CLI assembly",{"claudeCode":588},{"basePath":620,"githubOwner":593,"githubRepo":594,"locale":18,"slug":621,"type":248},"skills/ask","ask",{"evaluate":623,"extract":629},{"promptVersionExtension":205,"promptVersionScoring":206,"score":511,"tags":624,"targetMarket":218,"tier":277},[215,217,625,626,627,628],"llm","prompting","code-review","artifact-generation",{"commitSha":279,"license":280},{"parentExtensionId":606,"repoId":607},[628,217,215,627,625,626],{"evaluatedAt":633,"extractAt":611,"updatedAt":633},1778699303045,{"_creationTime":635,"_id":636,"community":637,"display":638,"identity":644,"providers":648,"relations":653,"tags":655,"workflow":656},1778697652123.8928,"k171pew5empzzrfghyg9nqrk6n86nqa9",{"reviewCount":8},{"description":639,"installMethods":640,"name":642,"sourceUrl":643},"Use gh for GitHub issues, PR status, CI/logs, comments, reviews, releases, and API queries.",{"claudeCode":641},"steipete/clawdis","github","https://github.com/steipete/clawdis",{"basePath":645,"githubOwner":646,"githubRepo":647,"locale":18,"slug":642,"type":248},"skills/github","steipete","clawdis",{"evaluate":649,"extract":652},{"promptVersionExtension":205,"promptVersionScoring":206,"score":511,"tags":650,"targetMarket":218,"tier":277},[642,215,542,651,217],"developer-tools",{"commitSha":279},{"repoId":654},"kd738npxg9yh3xf3vddzy9fyfh86nhng",[542,217,215,651,642],{"evaluatedAt":657,"extractAt":658,"updatedAt":657},1778698569289,1778697652123]