[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"extension-marketplace-firecrawl-cli-en":3,"guides-for-firecrawl-cli":463,"similar-k17be4khad1zr773wdg2m4yjhn86mjpr-en":464},{"_creationTime":4,"_id":5,"children":6,"community":47,"display":48,"evaluation":53,"identity":263,"isFallback":253,"parentExtension":265,"providers":266,"relations":273,"repo":274,"tags":461,"workflow":462},1778686940775.5703,"k17be4khad1zr773wdg2m4yjhn86mjpr",[7],{"_creationTime":8,"_id":9,"community":10,"display":12,"identity":17,"providers":22,"relations":41,"tags":43,"workflow":44},1778686940775.5706,"k17axfavjpz72cd3qqzn86shb186ncqt",{"reviewCount":11},0,{"description":13,"installMethods":14,"name":15,"sourceUrl":16},"Scrape, search, crawl, and map the web with a single command.",{"claudeCode":15},"firecrawl","https://github.com/firecrawl/cli",{"basePath":18,"githubOwner":15,"githubRepo":19,"locale":20,"slug":19,"type":21},"","cli","en","plugin",{"evaluate":23,"extract":36},{"promptVersionExtension":24,"promptVersionScoring":25,"score":26,"tags":27,"targetMarket":34,"tier":35},"3.0.0","4.4.0",78,[28,19,29,30,31,32,33],"web-scraping","automation","data-extraction","search","crawl","agent","global","community",{"commitSha":37,"plugin":38},"HEAD",{"mcpCount":11,"provider":39,"skillCount":40},"classify",9,{"parentExtensionId":5,"repoId":42},"kd7csd1wb06dg9c1jfy5063f2586ne60",[33,29,19,32,30,31,28],{"evaluatedAt":45,"extractAt":46,"updatedAt":45},1778686985594,1778686940775,{"reviewCount":11},{"description":49,"installMethods":50,"name":52,"sourceUrl":16},"CLI and Agent Skill for Firecrawl - Add scrape, search, and browsing capabilities to your AI agents",{"claudeCode":51},"firecrawl/cli","Firecrawl CLI",{"_creationTime":54,"_id":55,"extensionId":5,"locale":20,"result":56,"trustSignals":243,"workflow":261},1778686962954.9302,"kn71b1a0pgcbdr84r0hz891kwh86ny5t",{"checks":57,"evaluatedAt":204,"extensionSummary":205,"features":206,"nonGoals":213,"practices":217,"prerequisites":221,"promptVersionExtension":225,"promptVersionScoring":25,"purpose":226,"rationale":227,"score":228,"summary":229,"tags":230,"targetMarket":34,"tier":232,"useCases":233,"workflow":238},[58,63,66,69,73,76,80,84,87,90,94,98,101,105,108,111,114,117,120,123,127,131,136,140,144,147,150,154,158,161,164,167,170,173,176,180,184,188,191,195,198,201],{"category":59,"check":60,"severity":61,"summary":62},"Practical Utility","Problem relevance","pass","The description explicitly states that the CLI and Agent Skill for Firecrawl add scrape, search, and browsing capabilities to AI agents, directly addressing a user problem of extending agent functionality.",{"category":59,"check":64,"severity":61,"summary":65},"Unique selling proposition","The extension provides a comprehensive suite of tools (scrape, search, crawl, map, interact, agent) that significantly enhance AI agent capabilities beyond basic LLM interactions, offering specialized web data manipulation.",{"category":59,"check":67,"severity":61,"summary":68},"Production readiness","The CLI is fully implemented with numerous commands, options, and examples, supporting a complete workflow for web data extraction and agent integration.",{"category":70,"check":71,"severity":61,"summary":72},"Scope","Single responsibility principle","The extension focuses on web scraping, searching, and related agent capabilities, maintaining a coherent scope centered around web data interaction.",{"category":70,"check":74,"severity":61,"summary":75},"Description quality","The displayed description accurately and concisely reflects the extension's purpose of adding scraping, searching, and browsing to AI agents via its CLI and skills.",{"category":77,"check":78,"severity":61,"summary":79},"Invocation","Scoped tools","Tools are generally scoped as specific verb-noun actions (e.g., `scrape`, `search`, `map`, `crawl`), facilitating precise selection by the agent.",{"category":81,"check":82,"severity":61,"summary":83},"Documentation","Configuration & parameter reference","All command options, formats, and parameters are well-documented within the README, including descriptions and examples.",{"category":70,"check":85,"severity":61,"summary":86},"Tool naming","Tool names are descriptive and follow kebab-case, clearly indicating their function (e.g., `scrape`, `search`, `map`, `crawl`).",{"category":70,"check":88,"severity":61,"summary":89},"Minimal I/O surface","Command options and their descriptions suggest that inputs are specific and outputs are focused on the requested data formats, minimizing unnecessary data exposure.",{"category":91,"check":92,"severity":61,"summary":93},"License","License usability","The extension is licensed under the MIT license, a permissive open-source license, clearly indicated by the presence of a LICENSE file and the MIT SPDX identifier.",{"category":95,"check":96,"severity":61,"summary":97},"Maintenance","Commit recency","The repository shows recent commits, with the latest push occurring on May 12, 2026, indicating active maintenance.",{"category":95,"check":99,"severity":61,"summary":100},"Dependency Management","The presence of a lockfile (`package-lock.json` implied by `npm install -g`) suggests good dependency management practices.",{"category":102,"check":103,"severity":61,"summary":104},"Security","Secret Management","Secrets are handled via API keys or environment variables, with no hardcoded secrets found in the provided source code. The README clearly documents how to handle API keys.",{"category":102,"check":106,"severity":61,"summary":107},"Injection","The CLI commands are designed to take specific arguments and parameters, and there are no indications of loading or executing untrusted third-party code directly.",{"category":102,"check":109,"severity":61,"summary":110},"Transitive Supply-Chain Grenades","The CLI installs via npm and executes its own code, without evidence of runtime downloads of external scripts or instructions to follow fetched content.",{"category":102,"check":112,"severity":61,"summary":113},"Sandbox Isolation","The CLI operates on user-provided URLs and performs actions based on these inputs, without explicit evidence of attempting to modify files outside its intended scope or project folders.",{"category":102,"check":115,"severity":61,"summary":116},"Sandbox escape primitives","No evidence of detached processes or retry loops around denied tool calls was found in the provided script snippets.",{"category":102,"check":118,"severity":61,"summary":119},"Data Exfiltration","The extension handles API keys securely and its documentation does not suggest any undocumented outbound calls or exfiltration of confidential data.",{"category":102,"check":121,"severity":61,"summary":122},"Hidden Text Tricks","The README and provided code snippets appear to be free of hidden text tricks, invisible characters, or other obfuscation methods.",{"category":124,"check":125,"severity":61,"summary":126},"Hooks","Opaque code execution","The CLI is installed via npm and its scripts appear to be standard JavaScript, with no signs of obfuscation or runtime code fetching.",{"category":128,"check":129,"severity":61,"summary":130},"Portability","Structural Assumption","The CLI operates on URLs and can be configured via environment variables or CLI arguments, making it largely independent of specific project structures.",{"category":132,"check":133,"severity":134,"summary":135},"Trust","Issues Attention","info","In the last 90 days, 4 issues were opened and 1 was closed, indicating a closure rate below 50% and potentially slow response times.",{"category":137,"check":138,"severity":61,"summary":139},"Version","Release Management","The CLI version is clearly indicated in the `init` command (`@1.16.2`) and the `--version` global option, facilitating version management.",{"category":141,"check":142,"severity":61,"summary":143},"Code Execution","Validation","The CLI uses structured commands and options, and the README suggests that inputs are processed with appropriate validation for URLs, formats, and schemas.",{"category":102,"check":145,"severity":61,"summary":146},"Unguarded Destructive Operations","The primary operations are data retrieval and scraping, which are not inherently destructive. Commands like `cancel` are for active jobs and are guarded by job IDs.",{"category":141,"check":148,"severity":61,"summary":149},"Error Handling","The CLI provides clear error messages for authentication, configuration, and command execution, with options for JSON output to aid parsing.",{"category":141,"check":151,"severity":152,"summary":153},"Logging","not_applicable","This is a CLI tool, and its primary output is to stdout. Specific audit logging to a local file is not explicitly mentioned or required for this type of tool.",{"category":155,"check":156,"severity":152,"summary":157},"Compliance","GDPR","The tool scrapes public web content and handles API keys for authentication, but does not appear to operate on personal data directly without sanitization.",{"category":155,"check":159,"severity":61,"summary":160},"Target market","The extension is designed for general web scraping and agent integration, with no specific regional or jurisdictional limitations implied; it is global.",{"category":128,"check":162,"severity":61,"summary":163},"Runtime stability","As a Node.js CLI, it is expected to be cross-platform (Windows, macOS, Linux) and does not appear to rely on OS-specific features beyond standard Node.js APIs.",{"category":81,"check":165,"severity":61,"summary":166},"README","The README is comprehensive, clearly stating the extension's purpose, installation, authentication, and providing detailed command usage with examples.",{"category":70,"check":168,"severity":61,"summary":169},"Tool surface size","The CLI exposes a reasonable number of commands (scrape, search, map, crawl, agent, credit-usage, config, login/logout) suitable for its domain.",{"category":77,"check":171,"severity":61,"summary":172},"Overlapping near-synonym tools","The tools provided have distinct names and functionalities (e.g., `scrape` vs. `search` vs. `crawl`), avoiding significant overlap.",{"category":81,"check":174,"severity":61,"summary":175},"Phantom features","All advertised features, such as scraping, searching, crawling, and agent integration, have corresponding commands and detailed documentation in the README.",{"category":177,"check":178,"severity":61,"summary":179},"Install","Installation instruction","Installation instructions are clear and copy-pasteable, including options for global installation, agent setup, and authentication, with examples provided.",{"category":181,"check":182,"severity":61,"summary":183},"Errors","Actionable error messages","Error messages are generally informative, detailing what failed and providing guidance on authentication or configuration, with JSON output options for programmatic handling.",{"category":185,"check":186,"severity":61,"summary":187},"Execution","Pinned dependencies","The use of `npm install -g` and implied lockfile (`package-lock.json` for Node.js projects) indicates pinned dependencies for reproducible installations.",{"category":70,"check":189,"severity":152,"summary":190},"Dry-run preview","While some operations might be considered state-changing (like cancelling jobs), the core functions are data retrieval/processing, and explicit dry-run flags are not universally applicable or necessary for these.",{"category":192,"check":193,"severity":61,"summary":194},"Protocol","Idempotent retry & timeouts","The `timeout` option is available for many commands, and the nature of scraping/searching implies that retries on transient network issues would be beneficial, though explicit idempotency for mutations is less relevant here.",{"category":155,"check":196,"severity":61,"summary":197},"Telemetry opt-in","Telemetry is explicitly opt-in via the `FIRECRAWL_NO_TELEMETRY=1` environment variable, and the collected data is documented as anonymous usage data.",{"category":70,"check":199,"severity":61,"summary":200},"Theme declaration","The marketplace README declares a curation theme focused on AI agent capabilities for web scraping and data interaction, specifically for the Firecrawl ecosystem.",{"category":137,"check":202,"severity":152,"summary":203},"Per-entry version metadata","This is a CLI tool and not a marketplace itself, so per-entry version metadata for hosted plugins is not applicable.",1778686962686,"This is a command-line interface (CLI) and agent skill for Firecrawl that adds capabilities for scraping, searching, crawling, mapping, and interacting with web pages. It allows users to automate web data extraction and integrate these functionalities into AI agents.",[207,208,209,210,211,212],"Web scraping and content extraction","Web search and result aggregation","Website crawling and URL discovery","Interactive browser sessions","Agent skill integration","Structured data extraction via schema",[214,215,216],"Providing a visual web scraping UI","Replacing a full web browser for general user browsing","Managing complex application state beyond web scraping contexts",[218,219,220],"Web scraping best practices","Agent integration","Secure API key management",[222,223,224],"Node.js and npm/npx installed","Firecrawl API key (for cloud service) or self-hosted instance URL","An AI coding agent (for skill installation)","3.1.0","To empower AI agents and developers with powerful, integrated tools for comprehensive web data extraction and interaction directly from the terminal or via agent skills.","Excellent documentation and feature set, with active maintenance and clear security practices. Minor concern regarding issue closure rate.",97,"A robust and well-documented CLI tool for web scraping, search, and agent integration.",[28,19,231,30,31,32,29],"ai-agents","verified",[234,235,236,237],"Automating data collection from websites for analysis","Enriching AI agent capabilities with live web browsing","Performing targeted web searches and information retrieval","Developing applications that require programmatic web interaction",[239,240,241,242],"Install the CLI globally or via npx.","Authenticate with Firecrawl API key or via browser login.","Execute commands like `scrape`, `search`, `crawl`, or `agent` with appropriate arguments.","Optionally, install agent skills for AI agent integration using `firecrawl setup skills`.",{"codeQuality":244,"collectedAt":246,"documentation":247,"maintenance":250,"popularity":257,"security":259,"testCoverage":260},{"hasLockfile":245},true,1778686942474,{"descriptionLength":248,"readmeSize":249},99,25712,{"closedIssues90d":251,"forks":252,"hasChangelog":253,"openIssues90d":254,"pushedAt":255,"stars":256},1,49,false,4,1778599393000,383,{"npmDownloads":258},51144,{"hasNpmPackage":245,"smitheryVerified":253},{"hasCi":245,"hasTests":245},{"updatedAt":262},1778686962954,{"basePath":18,"githubOwner":15,"githubRepo":19,"locale":20,"slug":19,"type":264},"marketplace",null,{"evaluate":267,"extract":269},{"promptVersionExtension":225,"promptVersionScoring":25,"score":228,"tags":268,"targetMarket":34,"tier":232},[28,19,231,30,31,32,29],{"commitSha":37,"license":270,"marketplace":271,"plugin":272},"MIT",{"name":15,"pluginCount":251},{"mcpCount":11,"provider":39,"skillCount":11},{"repoId":42},{"_creationTime":275,"_id":42,"identity":276,"providers":277,"workflow":456},1778686934511.6108,{"githubOwner":15,"githubRepo":19,"sourceUrl":16},{"classify":278,"discover":448,"extract":451,"github":452,"npm":455},{"commitSha":37,"extensions":279},[280,290,316,326,333,341,349,357,365,373,381,389],{"basePath":18,"displayName":15,"installMethods":281,"rationale":282,"selectedPaths":283,"source":289,"sourceLanguage":20,"type":264},{"claudeCode":51},"marketplace.json at .claude-plugin/marketplace.json",[284,287],{"path":285,"priority":286},".claude-plugin/marketplace.json","mandatory",{"path":288,"priority":286},"README.md","rule",{"basePath":18,"description":13,"displayName":15,"installMethods":291,"rationale":292,"selectedPaths":293,"source":289,"sourceLanguage":20,"type":21},{"claudeCode":15},"plugin manifest at .claude-plugin/plugin.json",[294,296,297,300,302,304,306,308,310,312,314],{"path":295,"priority":286},".claude-plugin/plugin.json",{"path":288,"priority":286},{"path":298,"priority":299},"skills/firecrawl-agent/SKILL.md","medium",{"path":301,"priority":299},"skills/firecrawl-cli/SKILL.md",{"path":303,"priority":299},"skills/firecrawl-crawl/SKILL.md",{"path":305,"priority":299},"skills/firecrawl-download/SKILL.md",{"path":307,"priority":299},"skills/firecrawl-interact/SKILL.md",{"path":309,"priority":299},"skills/firecrawl-map/SKILL.md",{"path":311,"priority":299},"skills/firecrawl-parse/SKILL.md",{"path":313,"priority":299},"skills/firecrawl-scrape/SKILL.md",{"path":315,"priority":299},"skills/firecrawl-search/SKILL.md",{"basePath":317,"description":318,"displayName":319,"installMethods":320,"rationale":321,"selectedPaths":322,"source":289,"sourceLanguage":20,"type":325},"skills/firecrawl-agent","AI-powered autonomous data extraction that navigates complex sites and returns structured JSON. Use this skill when the user wants structured data from websites, needs to extract pricing tiers, product listings, directory entries, or any data as JSON with a schema. Triggers on \"extract structured data\", \"get all the products\", \"pull pricing info\", \"extract as JSON\", or when the user provides a JSON schema for website data. More powerful than simple scraping for multi-page structured extraction.\n","firecrawl-agent",{"claudeCode":51},"SKILL.md frontmatter at skills/firecrawl-agent/SKILL.md",[323],{"path":324,"priority":286},"SKILL.md","skill",{"basePath":327,"description":328,"displayName":15,"installMethods":329,"rationale":330,"selectedPaths":331,"source":289,"sourceLanguage":20,"type":325},"skills/firecrawl-cli","Search, scrape, and interact with the web via the Firecrawl CLI. Use this skill whenever the user wants to search the web, find articles, research a topic, look something up online, scrape a webpage, grab content from a URL, get data from a website, crawl documentation, download a site, or interact with pages that need clicks or logins. Also use when they say \"fetch this page\", \"pull the content from\", \"get the page at https://\", or reference external websites. This provides real-time web search with full page content and interact capabilities — beyond what Claude can do natively with built-in tools. Do NOT trigger for local file operations, git commands, deployments, or code editing tasks.\n",{"claudeCode":51},"SKILL.md frontmatter at skills/firecrawl-cli/SKILL.md",[332],{"path":324,"priority":286},{"basePath":334,"description":335,"displayName":336,"installMethods":337,"rationale":338,"selectedPaths":339,"source":289,"sourceLanguage":20,"type":325},"skills/firecrawl-crawl","Bulk extract content from an entire website or site section. Use this skill when the user wants to crawl a site, extract all pages from a docs section, bulk-scrape multiple pages following links, or says \"crawl\", \"get all the pages\", \"extract everything under /docs\", \"bulk extract\", or needs content from many pages on the same site. Handles depth limits, path filtering, and concurrent extraction.\n","firecrawl-crawl",{"claudeCode":51},"SKILL.md frontmatter at skills/firecrawl-crawl/SKILL.md",[340],{"path":324,"priority":286},{"basePath":342,"description":343,"displayName":344,"installMethods":345,"rationale":346,"selectedPaths":347,"source":289,"sourceLanguage":20,"type":325},"skills/firecrawl-download","Download an entire website as local files — markdown, screenshots, or multiple formats per page. Use this skill when the user wants to save a site locally, download documentation for offline use, bulk-save pages as files, or says \"download the site\", \"save as local files\", \"offline copy\", \"download all the docs\", or \"save for reference\". Combines site mapping and scraping into organized local directories.\n","firecrawl-download",{"claudeCode":51},"SKILL.md frontmatter at skills/firecrawl-download/SKILL.md",[348],{"path":324,"priority":286},{"basePath":350,"description":351,"displayName":352,"installMethods":353,"rationale":354,"selectedPaths":355,"source":289,"sourceLanguage":20,"type":325},"skills/firecrawl-interact","Control and interact with a live browser session on any scraped page — click buttons, fill forms, navigate flows, and extract data using natural language prompts or code. Use when the user needs to interact with a webpage beyond simple scraping: logging into a site, submitting forms, clicking through pagination, handling infinite scroll, navigating multi-step checkout or wizard flows, or when a regular scrape failed because content is behind JavaScript interaction. Also useful for authenticated scraping via profiles. Triggers on \"interact\", \"click\", \"fill out the form\", \"log in to\", \"sign in\", \"submit\", \"paginated\", \"next page\", \"infinite scroll\", \"interact with the page\", \"navigate to\", \"open a session\", or \"scrape failed\".\n","firecrawl-interact",{"claudeCode":51},"SKILL.md frontmatter at skills/firecrawl-interact/SKILL.md",[356],{"path":324,"priority":286},{"basePath":358,"description":359,"displayName":360,"installMethods":361,"rationale":362,"selectedPaths":363,"source":289,"sourceLanguage":20,"type":325},"skills/firecrawl-map","Discover and list all URLs on a website, with optional search filtering. Use this skill when the user wants to find a specific page on a large site, list all URLs, see the site structure, find where something is on a domain, or says \"map the site\", \"find the URL for\", \"what pages are on\", or \"list all pages\". Essential when the user knows which site but not which exact page.\n","firecrawl-map",{"claudeCode":51},"SKILL.md frontmatter at skills/firecrawl-map/SKILL.md",[364],{"path":324,"priority":286},{"basePath":366,"description":367,"displayName":368,"installMethods":369,"rationale":370,"selectedPaths":371,"source":289,"sourceLanguage":20,"type":325},"skills/firecrawl-parse","Efficiently extract and convert the contents of any local file—such as PDF, DOCX, DOC, ODT, RTF, XLSX, XLS, or HTML—into clean, well-formatted markdown saved to disk. Use this skill whenever the user requests to parse, read, or extract information from a file on their computer, including phrases like “parse this PDF”, “convert this document”, “read this file”, “extract text from”, or when a local file path (not a URL) is provided. This skill offers advanced options like generating AI-powered summaries and answering questions based on the file's content. Prefer this tool over `scrape` when handling local files to deliver precise, structured outputs for downstream tasks.\n","firecrawl-parse",{"claudeCode":51},"SKILL.md frontmatter at skills/firecrawl-parse/SKILL.md",[372],{"path":324,"priority":286},{"basePath":374,"description":375,"displayName":376,"installMethods":377,"rationale":378,"selectedPaths":379,"source":289,"sourceLanguage":20,"type":325},"skills/firecrawl-scrape","Extract clean markdown from any URL, including JavaScript-rendered SPAs. Use this skill whenever the user provides a URL and wants its content, says \"scrape\", \"grab\", \"fetch\", \"pull\", \"get the page\", \"extract from this URL\", or \"read this webpage\". Handles JS-rendered pages, multiple concurrent URLs, and returns LLM-optimized markdown. Use this instead of WebFetch for any webpage content extraction.\n","firecrawl-scrape",{"claudeCode":51},"SKILL.md frontmatter at skills/firecrawl-scrape/SKILL.md",[380],{"path":324,"priority":286},{"basePath":382,"description":383,"displayName":384,"installMethods":385,"rationale":386,"selectedPaths":387,"source":289,"sourceLanguage":20,"type":325},"skills/firecrawl-search","Web search with full page content extraction. Use this skill whenever the user asks to search the web, find articles, research a topic, look something up, find recent news, discover sources, or says \"search for\", \"find me\", \"look up\", \"what are people saying about\", or \"find articles about\". Returns real search results with optional full-page markdown — not just snippets. Provides capabilities beyond Claude's built-in WebSearch.\n","firecrawl-search",{"claudeCode":51},"SKILL.md frontmatter at skills/firecrawl-search/SKILL.md",[388],{"path":324,"priority":286},{"basePath":18,"description":390,"displayName":391,"installMethods":392,"license":393,"rationale":394,"selectedPaths":395,"source":289,"sourceLanguage":20,"type":19},"Command-line interface for Firecrawl. Scrape, crawl, and extract data from any website directly from your terminal.","firecrawl-cli",{"npm":391},"ISC","cli ecosystem detected at /",[396,398,399,402,404,406,408,410,412,414,416,418,420,422,424,426,428,430,432,434,436,438,440,442,444,446],{"path":397,"priority":286},"package.json",{"path":288,"priority":286},{"path":400,"priority":401},"src/index.ts","low",{"path":403,"priority":401},"src/commands/agent.ts",{"path":405,"priority":401},"src/commands/browser.ts",{"path":407,"priority":401},"src/commands/config.ts",{"path":409,"priority":401},"src/commands/crawl.ts",{"path":411,"priority":401},"src/commands/create.ts",{"path":413,"priority":401},"src/commands/credit-usage.ts",{"path":415,"priority":401},"src/commands/env.ts",{"path":417,"priority":401},"src/commands/experimental/backends.ts",{"path":419,"priority":401},"src/commands/experimental/index.ts",{"path":421,"priority":401},"src/commands/experimental/shared.ts",{"path":423,"priority":401},"src/commands/init.ts",{"path":425,"priority":401},"src/commands/interact.ts",{"path":427,"priority":401},"src/commands/login.ts",{"path":429,"priority":401},"src/commands/logout.ts",{"path":431,"priority":401},"src/commands/map.ts",{"path":433,"priority":401},"src/commands/parse.ts",{"path":435,"priority":401},"src/commands/scrape.ts",{"path":437,"priority":401},"src/commands/search.ts",{"path":439,"priority":401},"src/commands/setup.ts",{"path":441,"priority":401},"src/commands/skills-install.ts",{"path":443,"priority":401},"src/commands/skills-native.ts",{"path":445,"priority":401},"src/commands/status.ts",{"path":447,"priority":401},"src/commands/version.ts",{"sources":449},[450],"manual",{"npmPackage":391},{"closedIssues90d":251,"description":49,"forks":252,"homepage":453,"openIssues90d":254,"pushedAt":255,"readmeSize":249,"stars":256,"topics":454},"http://docs.firecrawl.dev/cli",[],{"downloads":258},{"classifiedAt":457,"discoverAt":458,"extractAt":459,"githubAt":459,"npmAt":460,"updatedAt":457},1778686940560,1778686934511,1778686936677,1778686938425,[231,29,19,32,30,31,28],{"evaluatedAt":262,"extractAt":46,"updatedAt":262},[],[465,497,528],{"_creationTime":466,"_id":467,"community":468,"display":469,"identity":475,"providers":477,"relations":490,"tags":492,"workflow":493},1778687361029.5278,"k17a6tjfs8x26pt762tjxft2zh86nnb6",{"reviewCount":11},{"description":470,"installMethods":471,"name":473,"sourceUrl":474},"SocratiCode — codebase intelligence plugin for Claude Code",{"claudeCode":472},"giancarloerra/socraticode","socraticode","https://github.com/giancarloerra/socraticode",{"basePath":18,"githubOwner":476,"githubRepo":473,"locale":20,"slug":473,"type":264},"giancarloerra",{"evaluate":478,"extract":487},{"promptVersionExtension":225,"promptVersionScoring":25,"score":479,"tags":480,"targetMarket":34,"tier":232},100,[481,482,483,31,484,485,486],"code-intelligence","codebase-analysis","ai-assistant","dependency-graph","local-docker","lsp",{"commitSha":37,"marketplace":488,"plugin":489},{"name":473,"pluginCount":251},{"mcpCount":11,"provider":39,"skillCount":11},{"repoId":491},"kd76w3r9kxfskg0y8kxc68drnd86m8rk",[483,481,482,484,485,486,31],{"evaluatedAt":494,"extractAt":495,"updatedAt":496},1778687383326,1778687361029,1778687533262,{"_creationTime":498,"_id":499,"community":500,"display":501,"identity":507,"providers":510,"relations":522,"tags":524,"workflow":525},1778699018122.7656,"k17cywe30jfsfw3cdpncjfn8y186nvyw",{"reviewCount":11},{"description":502,"installMethods":503,"name":505,"sourceUrl":506},"Production-ready workflow orchestration with 80 focused plugins, 185 specialized agents, and 153 skills - optimized for granular installation and minimal token usage",{"claudeCode":504},"wshobson/agents","Claude Code Plugins","https://github.com/wshobson/agents",{"basePath":18,"githubOwner":508,"githubRepo":509,"locale":20,"slug":509,"type":264},"wshobson","agents",{"evaluate":511,"extract":517},{"promptVersionExtension":225,"promptVersionScoring":25,"score":248,"tags":512,"targetMarket":34,"tier":232},[513,514,231,515,29,516],"workflow-orchestration","developer-tools","plugins","code-generation",{"commitSha":37,"license":270,"marketplace":518,"plugin":521},{"name":519,"pluginCount":520},"claude-code-workflows",81,{"mcpCount":11,"provider":39,"skillCount":11},{"repoId":523},"kd74de64zj0axtg5b8t7eqqe2x86nske",[231,29,516,514,515,513],{"evaluatedAt":526,"extractAt":527,"updatedAt":526},1778699049186,1778699018122,{"_creationTime":529,"_id":530,"community":531,"display":532,"identity":538,"providers":541,"relations":552,"tags":554,"workflow":555},1778686640222.7908,"k17ef8php9wk308mkg59kdp6b186nzhx",{"reviewCount":11},{"description":533,"installMethods":534,"name":536,"sourceUrl":537},"Cursor plugin marketplace for Every Inc plugins",{"claudeCode":535},"EveryInc/compound-engineering-plugin","Compound Engineering","https://github.com/EveryInc/compound-engineering-plugin",{"basePath":18,"githubOwner":539,"githubRepo":540,"locale":20,"slug":540,"type":264},"EveryInc","compound-engineering-plugin",{"evaluate":542,"extract":548},{"promptVersionExtension":225,"promptVersionScoring":25,"score":248,"tags":543,"targetMarket":34,"tier":232},[544,545,546,547,231],"development-workflow","code-review","planning","documentation",{"commitSha":37,"license":270,"marketplace":549,"plugin":551},{"name":540,"pluginCount":550},2,{"mcpCount":11,"provider":39,"skillCount":11},{"repoId":553},"kd7e40my1b5g70tg0f60qg85ch86nn08",[231,545,544,547,546],{"evaluatedAt":556,"extractAt":557,"updatedAt":556},1778686722515,1778686640222]