[{"data":1,"prerenderedAt":-1},["ShallowReactive",2],{"extension-skill-firecrawl-firecrawl-map-de":3,"guides-for-firecrawl-firecrawl-map":493,"similar-k17c5kkgjkbd4f35bh8jv5js9h86n4gc-de":494},{"_creationTime":4,"_id":5,"children":6,"community":7,"display":9,"evaluation":15,"identity":244,"isFallback":226,"parentExtension":248,"providers":303,"relations":307,"repo":308,"tags":491,"workflow":492},1778686940775.572,"k17c5kkgjkbd4f35bh8jv5js9h86n4gc",[],{"reviewCount":8},0,{"description":10,"installMethods":11,"name":13,"sourceUrl":14},"Discover and list all URLs on a website, with optional search filtering. Use this skill when the user wants to find a specific page on a large site, list all URLs, see the site structure, find where something is on a domain, or says \"map the site\", \"find the URL for\", \"what pages are on\", or \"list all pages\". Essential when the user knows which site but not which exact page.\n",{"claudeCode":12},"firecrawl/cli","firecrawl-map","https://github.com/firecrawl/cli",{"_creationTime":16,"_id":17,"extensionId":5,"locale":18,"result":19,"trustSignals":224,"workflow":242},1778687127164.1436,"kn77shb1ga5qkwvwssfpamc17h86nq5t","en",{"checks":20,"evaluatedAt":193,"extensionSummary":194,"features":195,"nonGoals":201,"promptVersionExtension":205,"promptVersionScoring":206,"purpose":207,"rationale":208,"score":209,"summary":210,"tags":211,"targetMarket":217,"tier":218,"useCases":219},[21,26,29,32,36,39,43,47,50,53,57,61,64,68,71,74,77,80,83,86,90,94,99,103,107,110,114,118,122,125,128,131,134,137,140,144,148,151,154,158,161,164,167,170,174,177,180,183,186,190],{"category":22,"check":23,"severity":24,"summary":25},"Practical Utility","Problem relevance","pass","The description clearly states the problem of finding specific URLs on large websites and lists several relevant user intents and scenarios.",{"category":22,"check":27,"severity":24,"summary":28},"Unique selling proposition","The skill offers a focused capability for discovering URLs, which is a distinct and useful function beyond general web scraping or searching.",{"category":22,"check":30,"severity":24,"summary":31},"Production readiness","The tool is production-ready, offering a clear command-line interface for discovering URLs with various options for filtering and output.",{"category":33,"check":34,"severity":24,"summary":35},"Scope","Single responsibility principle","The extension is focused solely on discovering and listing URLs on a website, adhering to a single responsibility.",{"category":33,"check":37,"severity":24,"summary":38},"Description quality","The displayed description accurately reflects the functionality of discovering and listing URLs on a website with optional search filtering.",{"category":40,"check":41,"severity":24,"summary":42},"Invocation","Scoped tools","The tool exposes a single, well-defined command 'map' for its specific purpose.",{"category":44,"check":45,"severity":24,"summary":46},"Documentation","Configuration & parameter reference","All options for the `map` command, including limits, search, sitemap handling, subdomain inclusion, JSON output, and file output, are clearly documented with descriptions.",{"category":33,"check":48,"severity":24,"summary":49},"Tool naming","The tool name 'map' is descriptive and aligns with its function of discovering URLs on a website.",{"category":33,"check":51,"severity":24,"summary":52},"Minimal I/O surface","The input parameters for the 'map' command are specific and well-defined, and the output options (plain text or JSON) are clear.",{"category":54,"check":55,"severity":24,"summary":56},"License","License usability","The project is licensed under the MIT license, which is a permissive open-source license.",{"category":58,"check":59,"severity":24,"summary":60},"Maintenance","Commit recency","The last commit was on May 12, 2026, indicating recent maintenance.",{"category":58,"check":62,"severity":24,"summary":63},"Dependency Management","The project uses npm and has a lockfile, indicating managed dependencies.",{"category":65,"check":66,"severity":24,"summary":67},"Security","Secret Management","The extension does not appear to handle secrets directly; authentication uses API keys that are managed via environment variables or a secure login process.",{"category":65,"check":69,"severity":24,"summary":70},"Injection","The tool focuses on URL discovery and does not appear to execute arbitrary third-party code or instructions.",{"category":65,"check":72,"severity":24,"summary":73},"Transitive Supply-Chain Grenades","The extension operates on provided URLs and does not fetch or execute external code at runtime.",{"category":65,"check":75,"severity":24,"summary":76},"Sandbox Isolation","The CLI tool operates within its own scope and does not appear to modify files outside of specified output paths.",{"category":65,"check":78,"severity":24,"summary":79},"Sandbox escape primitives","No evidence of detached process spawns or deny-retry loops was found in the provided code.",{"category":65,"check":81,"severity":24,"summary":82},"Data Exfiltration","The tool's primary function is URL discovery, and there are no indications of exfiltrating confidential data.",{"category":65,"check":84,"severity":24,"summary":85},"Hidden Text Tricks","The bundled content and documentation appear to be free of hidden steering tricks.",{"category":87,"check":88,"severity":24,"summary":89},"Hooks","Opaque code execution","The CLI commands are written in standard TypeScript/Node.js and are not obfuscated.",{"category":91,"check":92,"severity":24,"summary":93},"Portability","Structural Assumption","The tool makes no assumptions about user project structure, only requiring a URL as input.",{"category":95,"check":96,"severity":97,"summary":98},"Trust","Issues Attention","info","There were 4 issues opened and 1 closed in the last 90 days, indicating slow response times.",{"category":100,"check":101,"severity":24,"summary":102},"Versioning","Release Management","The CLI reports version 1.16.2 in its status output, indicating a versioned release.",{"category":104,"check":105,"severity":24,"summary":106},"Execution","Validation","Input arguments like URLs and file paths are likely validated by the underlying libraries used in the CLI.",{"category":65,"check":108,"severity":24,"summary":109},"Unguarded Destructive Operations","The 'map' command is read-only and does not perform destructive operations.",{"category":111,"check":112,"severity":24,"summary":113},"Code Execution","Error Handling","The CLI provides structured output and likely handles errors gracefully, providing informative messages.",{"category":111,"check":115,"severity":116,"summary":117},"Logging","not_applicable","The tool is primarily a CLI for data extraction and does not involve destructive actions or outbound calls that would typically require a local audit log.",{"category":119,"check":120,"severity":116,"summary":121},"Compliance","GDPR","The tool's function is to discover public URLs and does not process personal data.",{"category":119,"check":123,"severity":24,"summary":124},"Target market","The extension operates on publicly accessible websites and has no regional limitations.",{"category":91,"check":126,"severity":24,"summary":127},"Runtime stability","As a Node.js CLI tool, it is expected to run on standard POSIX-compliant systems.",{"category":44,"check":129,"severity":24,"summary":130},"README","The README file is comprehensive, providing installation, authentication, and usage instructions for various commands.",{"category":33,"check":132,"severity":24,"summary":133},"Tool surface size","The 'map' command is one of several commands within the CLI, keeping the overall tool surface manageable.",{"category":40,"check":135,"severity":24,"summary":136},"Overlapping near-synonym tools","The 'map' command is distinct and does not have near-synonyms within the Firecrawl CLI.",{"category":44,"check":138,"severity":24,"summary":139},"Phantom features","All advertised features for the 'map' command are implemented and documented.",{"category":141,"check":142,"severity":24,"summary":143},"Install","Installation instruction","The README provides clear `npm install -g` instructions and copy-pasteable examples for using the 'map' command.",{"category":145,"check":146,"severity":24,"summary":147},"Errors","Actionable error messages","The CLI is expected to provide actionable error messages, given its professional implementation.",{"category":104,"check":149,"severity":24,"summary":150},"Pinned dependencies","The project utilizes npm with a lockfile, indicating pinned dependencies.",{"category":33,"check":152,"severity":24,"summary":153},"Dry-run preview","The 'map' command is read-only and does not require a dry-run feature.",{"category":155,"check":156,"severity":116,"summary":157},"Protocol","Idempotent retry & timeouts","The 'map' command is a read-only operation and does not involve state-changing operations or remote calls requiring idempotency.",{"category":119,"check":159,"severity":24,"summary":160},"Telemetry opt-in","Telemetry is opt-in via an environment variable, and the collected data is documented as anonymous usage statistics.",{"category":40,"check":162,"severity":24,"summary":163},"Precise Purpose","The skill's purpose is precisely defined, stating it discovers and lists URLs on a website with optional search filtering, and provides clear usage scenarios.",{"category":40,"check":165,"severity":24,"summary":166},"Concise Frontmatter","The frontmatter is concise, clearly stating the core capability and providing trigger phrases within a reasonable character limit.",{"category":44,"check":168,"severity":24,"summary":169},"Concise Body","The SKILL.md is concise, with detailed options and examples provided within the main document, and references to other commands.",{"category":171,"check":172,"severity":24,"summary":173},"Context","Progressive Disclosure","The SKILL.md outlines the 'map' command's functionality and options directly, without needing external files for this specific command.",{"category":171,"check":175,"severity":116,"summary":176},"Forked exploration","This skill is a focused utility for URL discovery and does not involve deep exploration requiring a forked context.",{"category":22,"check":178,"severity":24,"summary":179},"Usage examples","Clear, ready-to-use examples are provided in the SKILL.md and README for various use cases of the 'map' command.",{"category":22,"check":181,"severity":24,"summary":182},"Edge cases","The documentation implicitly handles edge cases by providing options for sitemap handling, subdomain inclusion, and limiting results.",{"category":111,"check":184,"severity":116,"summary":185},"Tool Fallback","This extension is a standalone CLI tool and does not rely on external MCP servers or other skills, thus no fallback is needed.",{"category":187,"check":188,"severity":24,"summary":189},"Safety","Halt on unexpected state","As a read-only tool, it does not have destructive operations and thus does not require specific pre-state checks for halting.",{"category":91,"check":191,"severity":24,"summary":192},"Cross-skill coupling","The 'map' skill operates independently and does not implicitly rely on other skills.",1778687127057,"This is a command-line interface tool that discovers and lists all URLs on a given website. It can filter results by search queries, handle sitemaps, and include subdomains.",[196,197,198,199,200],"Discover all URLs on a website","Filter URLs by search query","Include/exclude subdomains","Handle sitemaps","Output in plain text or JSON",[202,203,204],"Scraping the content of discovered URLs (use `firecrawl-scrape` for that)","Crawling an entire website structure (use `firecrawl-crawl` for deep crawling)","Executing arbitrary commands or code","3.0.0","4.4.0","To efficiently find and list all URLs on a website, making it easier to understand site structure or target specific pages for further action.","The extension is well-documented, production-ready, and focuses on a single, clear responsibility with excellent usage examples. The only minor point is the slow issue closure rate.",97,"A robust CLI tool for discovering and listing website URLs with flexible filtering.",[212,213,214,215,216],"web","scraping","url","cli","discovery","global","verified",[220,221,222,223],"Finding specific pages on large websites","Listing all URLs for site mapping or analysis","Identifying where specific content resides on a domain","Preparing for bulk scraping or crawling by identifying target URLs",{"codeQuality":225,"collectedAt":227,"documentation":228,"maintenance":231,"popularity":238,"security":240,"testCoverage":241},{"hasLockfile":226},true,1778687116421,{"descriptionLength":229,"readmeSize":230},378,25712,{"closedIssues90d":232,"forks":233,"hasChangelog":234,"openIssues90d":235,"pushedAt":236,"stars":237},1,49,false,4,1778599393000,383,{"npmDownloads":239},51144,{"hasNpmPackage":226,"smitheryVerified":234},{"hasCi":226,"hasTests":226},{"updatedAt":243},1778687127164,{"basePath":245,"githubOwner":246,"githubRepo":215,"locale":18,"slug":13,"type":247},"skills/firecrawl-map","firecrawl","skill",{"_creationTime":249,"_id":250,"community":251,"display":252,"identity":255,"parentExtension":258,"providers":290,"relations":299,"tags":300,"workflow":301},1778686940775.5706,"k17axfavjpz72cd3qqzn86shb186ncqt",{"reviewCount":8},{"description":253,"installMethods":254,"name":246,"sourceUrl":14},"Scrape, search, crawl, and map the web with a single command.",{"claudeCode":246},{"basePath":256,"githubOwner":246,"githubRepo":215,"locale":18,"slug":215,"type":257},"","plugin",{"_creationTime":259,"_id":260,"community":261,"display":262,"identity":266,"providers":268,"relations":284,"tags":286,"workflow":287},1778686940775.5703,"k17be4khad1zr773wdg2m4yjhn86mjpr",{"reviewCount":8},{"description":263,"installMethods":264,"name":265,"sourceUrl":14},"CLI and Agent Skill for Firecrawl - Add scrape, search, and browsing capabilities to your AI agents",{"claudeCode":12},"Firecrawl CLI",{"basePath":256,"githubOwner":246,"githubRepo":215,"locale":18,"slug":215,"type":267},"marketplace",{"evaluate":269,"extract":278},{"promptVersionExtension":270,"promptVersionScoring":206,"score":209,"tags":271,"targetMarket":217,"tier":218},"3.1.0",[272,215,273,274,275,276,277],"web-scraping","ai-agents","data-extraction","search","crawl","automation",{"commitSha":279,"license":280,"marketplace":281,"plugin":282},"HEAD","MIT",{"name":246,"pluginCount":232},{"mcpCount":8,"provider":283,"skillCount":8},"classify",{"repoId":285},"kd7csd1wb06dg9c1jfy5063f2586ne60",[273,277,215,276,274,275,272],{"evaluatedAt":288,"extractAt":289,"updatedAt":288},1778686962954,1778686940775,{"evaluate":291,"extract":296},{"promptVersionExtension":205,"promptVersionScoring":206,"score":292,"tags":293,"targetMarket":217,"tier":295},78,[272,215,277,274,275,276,294],"agent","community",{"commitSha":279,"plugin":297},{"mcpCount":8,"provider":283,"skillCount":298},9,{"parentExtensionId":260,"repoId":285},[294,277,215,276,274,275,272],{"evaluatedAt":302,"extractAt":289,"updatedAt":302},1778686985594,{"evaluate":304,"extract":306},{"promptVersionExtension":205,"promptVersionScoring":206,"score":209,"tags":305,"targetMarket":217,"tier":218},[212,213,214,215,216],{"commitSha":279},{"parentExtensionId":250,"repoId":285},{"_creationTime":309,"_id":285,"identity":310,"providers":311,"workflow":486},1778686934511.6108,{"githubOwner":246,"githubRepo":215,"sourceUrl":14},{"classify":312,"discover":478,"extract":481,"github":482,"npm":485},{"commitSha":279,"extensions":313},[314,324,350,359,366,374,382,390,395,403,411,419],{"basePath":256,"displayName":246,"installMethods":315,"rationale":316,"selectedPaths":317,"source":323,"sourceLanguage":18,"type":267},{"claudeCode":12},"marketplace.json at .claude-plugin/marketplace.json",[318,321],{"path":319,"priority":320},".claude-plugin/marketplace.json","mandatory",{"path":322,"priority":320},"README.md","rule",{"basePath":256,"description":253,"displayName":246,"installMethods":325,"rationale":326,"selectedPaths":327,"source":323,"sourceLanguage":18,"type":257},{"claudeCode":246},"plugin manifest at .claude-plugin/plugin.json",[328,330,331,334,336,338,340,342,344,346,348],{"path":329,"priority":320},".claude-plugin/plugin.json",{"path":322,"priority":320},{"path":332,"priority":333},"skills/firecrawl-agent/SKILL.md","medium",{"path":335,"priority":333},"skills/firecrawl-cli/SKILL.md",{"path":337,"priority":333},"skills/firecrawl-crawl/SKILL.md",{"path":339,"priority":333},"skills/firecrawl-download/SKILL.md",{"path":341,"priority":333},"skills/firecrawl-interact/SKILL.md",{"path":343,"priority":333},"skills/firecrawl-map/SKILL.md",{"path":345,"priority":333},"skills/firecrawl-parse/SKILL.md",{"path":347,"priority":333},"skills/firecrawl-scrape/SKILL.md",{"path":349,"priority":333},"skills/firecrawl-search/SKILL.md",{"basePath":351,"description":352,"displayName":353,"installMethods":354,"rationale":355,"selectedPaths":356,"source":323,"sourceLanguage":18,"type":247},"skills/firecrawl-agent","AI-powered autonomous data extraction that navigates complex sites and returns structured JSON. Use this skill when the user wants structured data from websites, needs to extract pricing tiers, product listings, directory entries, or any data as JSON with a schema. Triggers on \"extract structured data\", \"get all the products\", \"pull pricing info\", \"extract as JSON\", or when the user provides a JSON schema for website data. More powerful than simple scraping for multi-page structured extraction.\n","firecrawl-agent",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-agent/SKILL.md",[357],{"path":358,"priority":320},"SKILL.md",{"basePath":360,"description":361,"displayName":246,"installMethods":362,"rationale":363,"selectedPaths":364,"source":323,"sourceLanguage":18,"type":247},"skills/firecrawl-cli","Search, scrape, and interact with the web via the Firecrawl CLI. Use this skill whenever the user wants to search the web, find articles, research a topic, look something up online, scrape a webpage, grab content from a URL, get data from a website, crawl documentation, download a site, or interact with pages that need clicks or logins. Also use when they say \"fetch this page\", \"pull the content from\", \"get the page at https://\", or reference external websites. This provides real-time web search with full page content and interact capabilities — beyond what Claude can do natively with built-in tools. Do NOT trigger for local file operations, git commands, deployments, or code editing tasks.\n",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-cli/SKILL.md",[365],{"path":358,"priority":320},{"basePath":367,"description":368,"displayName":369,"installMethods":370,"rationale":371,"selectedPaths":372,"source":323,"sourceLanguage":18,"type":247},"skills/firecrawl-crawl","Bulk extract content from an entire website or site section. Use this skill when the user wants to crawl a site, extract all pages from a docs section, bulk-scrape multiple pages following links, or says \"crawl\", \"get all the pages\", \"extract everything under /docs\", \"bulk extract\", or needs content from many pages on the same site. Handles depth limits, path filtering, and concurrent extraction.\n","firecrawl-crawl",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-crawl/SKILL.md",[373],{"path":358,"priority":320},{"basePath":375,"description":376,"displayName":377,"installMethods":378,"rationale":379,"selectedPaths":380,"source":323,"sourceLanguage":18,"type":247},"skills/firecrawl-download","Download an entire website as local files — markdown, screenshots, or multiple formats per page. Use this skill when the user wants to save a site locally, download documentation for offline use, bulk-save pages as files, or says \"download the site\", \"save as local files\", \"offline copy\", \"download all the docs\", or \"save for reference\". Combines site mapping and scraping into organized local directories.\n","firecrawl-download",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-download/SKILL.md",[381],{"path":358,"priority":320},{"basePath":383,"description":384,"displayName":385,"installMethods":386,"rationale":387,"selectedPaths":388,"source":323,"sourceLanguage":18,"type":247},"skills/firecrawl-interact","Control and interact with a live browser session on any scraped page — click buttons, fill forms, navigate flows, and extract data using natural language prompts or code. Use when the user needs to interact with a webpage beyond simple scraping: logging into a site, submitting forms, clicking through pagination, handling infinite scroll, navigating multi-step checkout or wizard flows, or when a regular scrape failed because content is behind JavaScript interaction. Also useful for authenticated scraping via profiles. Triggers on \"interact\", \"click\", \"fill out the form\", \"log in to\", \"sign in\", \"submit\", \"paginated\", \"next page\", \"infinite scroll\", \"interact with the page\", \"navigate to\", \"open a session\", or \"scrape failed\".\n","firecrawl-interact",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-interact/SKILL.md",[389],{"path":358,"priority":320},{"basePath":245,"description":10,"displayName":13,"installMethods":391,"rationale":392,"selectedPaths":393,"source":323,"sourceLanguage":18,"type":247},{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-map/SKILL.md",[394],{"path":358,"priority":320},{"basePath":396,"description":397,"displayName":398,"installMethods":399,"rationale":400,"selectedPaths":401,"source":323,"sourceLanguage":18,"type":247},"skills/firecrawl-parse","Efficiently extract and convert the contents of any local file—such as PDF, DOCX, DOC, ODT, RTF, XLSX, XLS, or HTML—into clean, well-formatted markdown saved to disk. Use this skill whenever the user requests to parse, read, or extract information from a file on their computer, including phrases like “parse this PDF”, “convert this document”, “read this file”, “extract text from”, or when a local file path (not a URL) is provided. This skill offers advanced options like generating AI-powered summaries and answering questions based on the file's content. Prefer this tool over `scrape` when handling local files to deliver precise, structured outputs for downstream tasks.\n","firecrawl-parse",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-parse/SKILL.md",[402],{"path":358,"priority":320},{"basePath":404,"description":405,"displayName":406,"installMethods":407,"rationale":408,"selectedPaths":409,"source":323,"sourceLanguage":18,"type":247},"skills/firecrawl-scrape","Extract clean markdown from any URL, including JavaScript-rendered SPAs. Use this skill whenever the user provides a URL and wants its content, says \"scrape\", \"grab\", \"fetch\", \"pull\", \"get the page\", \"extract from this URL\", or \"read this webpage\". Handles JS-rendered pages, multiple concurrent URLs, and returns LLM-optimized markdown. Use this instead of WebFetch for any webpage content extraction.\n","firecrawl-scrape",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-scrape/SKILL.md",[410],{"path":358,"priority":320},{"basePath":412,"description":413,"displayName":414,"installMethods":415,"rationale":416,"selectedPaths":417,"source":323,"sourceLanguage":18,"type":247},"skills/firecrawl-search","Web search with full page content extraction. Use this skill whenever the user asks to search the web, find articles, research a topic, look something up, find recent news, discover sources, or says \"search for\", \"find me\", \"look up\", \"what are people saying about\", or \"find articles about\". Returns real search results with optional full-page markdown — not just snippets. Provides capabilities beyond Claude's built-in WebSearch.\n","firecrawl-search",{"claudeCode":12},"SKILL.md frontmatter at skills/firecrawl-search/SKILL.md",[418],{"path":358,"priority":320},{"basePath":256,"description":420,"displayName":421,"installMethods":422,"license":423,"rationale":424,"selectedPaths":425,"source":323,"sourceLanguage":18,"type":215},"Command-line interface for Firecrawl. Scrape, crawl, and extract data from any website directly from your terminal.","firecrawl-cli",{"npm":421},"ISC","cli ecosystem detected at /",[426,428,429,432,434,436,438,440,442,444,446,448,450,452,454,456,458,460,462,464,466,468,470,472,474,476],{"path":427,"priority":320},"package.json",{"path":322,"priority":320},{"path":430,"priority":431},"src/index.ts","low",{"path":433,"priority":431},"src/commands/agent.ts",{"path":435,"priority":431},"src/commands/browser.ts",{"path":437,"priority":431},"src/commands/config.ts",{"path":439,"priority":431},"src/commands/crawl.ts",{"path":441,"priority":431},"src/commands/create.ts",{"path":443,"priority":431},"src/commands/credit-usage.ts",{"path":445,"priority":431},"src/commands/env.ts",{"path":447,"priority":431},"src/commands/experimental/backends.ts",{"path":449,"priority":431},"src/commands/experimental/index.ts",{"path":451,"priority":431},"src/commands/experimental/shared.ts",{"path":453,"priority":431},"src/commands/init.ts",{"path":455,"priority":431},"src/commands/interact.ts",{"path":457,"priority":431},"src/commands/login.ts",{"path":459,"priority":431},"src/commands/logout.ts",{"path":461,"priority":431},"src/commands/map.ts",{"path":463,"priority":431},"src/commands/parse.ts",{"path":465,"priority":431},"src/commands/scrape.ts",{"path":467,"priority":431},"src/commands/search.ts",{"path":469,"priority":431},"src/commands/setup.ts",{"path":471,"priority":431},"src/commands/skills-install.ts",{"path":473,"priority":431},"src/commands/skills-native.ts",{"path":475,"priority":431},"src/commands/status.ts",{"path":477,"priority":431},"src/commands/version.ts",{"sources":479},[480],"manual",{"npmPackage":421},{"closedIssues90d":232,"description":263,"forks":233,"homepage":483,"openIssues90d":235,"pushedAt":236,"readmeSize":230,"stars":237,"topics":484},"http://docs.firecrawl.dev/cli",[],{"downloads":239},{"classifiedAt":487,"discoverAt":488,"extractAt":489,"githubAt":489,"npmAt":490,"updatedAt":487},1778686940560,1778686934511,1778686936677,1778686938425,[215,216,213,214,212],{"evaluatedAt":243,"extractAt":289,"updatedAt":243},[],[495,524,553,584,612,636],{"_creationTime":496,"_id":497,"community":498,"display":499,"identity":505,"providers":509,"relations":517,"tags":520,"workflow":521},1778695548458.3691,"k170v9zdk5jd1g1dwtya1fgzfx86n6dm",{"reviewCount":8},{"description":500,"installMethods":501,"name":503,"sourceUrl":504},"Install skills, agents, and teams from agent-almanac into any supported agentic framework using the CLI. Covers framework detection, content search, installation with dependency resolution, health auditing, and manifest-based syncing. Use when setting up a new project with agentic capabilities, installing specific skills or entire domains, targeting multiple frameworks simultaneously, or maintaining a declarative manifest of installed content.\n",{"claudeCode":502},"pjt222/agent-almanac","install-almanac-content","https://github.com/pjt222/agent-almanac",{"basePath":506,"githubOwner":507,"githubRepo":508,"locale":18,"slug":503,"type":247},"skills/install-almanac-content","pjt222","agent-almanac",{"evaluate":510,"extract":516},{"promptVersionExtension":205,"promptVersionScoring":206,"score":511,"tags":512,"targetMarket":217,"tier":218},100,[215,513,514,216,515],"installation","framework-integration","agent-management",{"commitSha":279},{"parentExtensionId":518,"repoId":519},"k170h0janaa9kwn7cfgfz2ykss86mmh9","kd7aryv63z61j39n2td1aeqkvh86mh12",[515,215,216,514,513],{"evaluatedAt":522,"extractAt":523,"updatedAt":522},1778698810606,1778695548458,{"_creationTime":525,"_id":526,"community":527,"display":528,"identity":534,"providers":538,"relations":546,"tags":549,"workflow":550},1778683460321.3828,"k178byhs7vswk8154tmztnehgx86n47n",{"reviewCount":8},{"description":529,"installMethods":530,"name":532,"sourceUrl":533},"Use this skill when the user wants to retrieve a URL without a full browser session: fetch HTML or JSON from static pages, inspect status codes or headers, follow redirects, or get page source for simple scraping. Prefer it over a browser when JavaScript rendering and page interaction are not needed. Supports proxies and redirect control.",{"claudeCode":531},"browserbase/skills","fetch","https://github.com/browserbase/skills",{"basePath":535,"githubOwner":536,"githubRepo":537,"locale":18,"slug":532,"type":247},"skills/fetch","browserbase","skills",{"evaluate":539,"extract":545},{"promptVersionExtension":205,"promptVersionScoring":206,"score":540,"tags":541,"targetMarket":217,"tier":218},98,[212,542,213,543,544],"http","api","data-retrieval",{"commitSha":279},{"parentExtensionId":547,"repoId":548},"k17bx77jb71yrwatep2nb43r0d86m6cb","kd77wvcdm5fq9xp8hk6ppm832s86myxr",[543,544,542,213,212],{"evaluatedAt":551,"extractAt":552,"updatedAt":551},1778683789383,1778683460321,{"_creationTime":554,"_id":555,"community":556,"display":557,"identity":563,"providers":568,"relations":577,"tags":580,"workflow":581},1778696691708.3274,"k170az7r02e9e2v47mpy80kx6n86nff3",{"reviewCount":8},{"description":558,"installMethods":559,"name":561,"sourceUrl":562},"Detect current market regime using npx neural-trader — bull/bear/ranging/volatile classification with recommended strategy",{"claudeCode":560},"ruvnet/ruflo","Trader Regime","https://github.com/ruvnet/ruflo",{"basePath":564,"githubOwner":565,"githubRepo":566,"locale":18,"slug":567,"type":247},"plugins/ruflo-neural-trader/skills/trader-regime","ruvnet","ruflo","trader-regime",{"evaluate":569,"extract":576},{"promptVersionExtension":205,"promptVersionScoring":206,"score":511,"tags":570,"targetMarket":217,"tier":218},[571,572,573,574,575,215],"finance","trading","market-analysis","ai","typescript",{"commitSha":279,"license":280},{"parentExtensionId":578,"repoId":579},"k17drge8h1fgzchr0p4jaeg33n86mwmy","kd7ed28gj8n0y3msk5dzrp05zs86nqtc",[574,215,571,573,572,575],{"evaluatedAt":582,"extractAt":583,"updatedAt":582},1778701108877,1778696691708,{"_creationTime":585,"_id":586,"community":587,"display":588,"identity":594,"providers":598,"relations":605,"tags":608,"workflow":609},1778699234184.6174,"k174zww66m804nhr89ttra7r6d86nwyg",{"reviewCount":8},{"description":589,"installMethods":590,"name":592,"sourceUrl":593},"Use first for install/update routing — sends setup, doctor, or MCP requests to the correct OMC setup flow",{"claudeCode":591},"Yeachan-Heo/oh-my-claudecode","setup","https://github.com/Yeachan-Heo/oh-my-claudecode",{"basePath":595,"githubOwner":596,"githubRepo":597,"locale":18,"slug":592,"type":247},"skills/setup","Yeachan-Heo","oh-my-claudecode",{"evaluate":599,"extract":604},{"promptVersionExtension":205,"promptVersionScoring":206,"score":511,"tags":600,"targetMarket":217,"tier":218},[592,601,602,215,603],"routing","configuration","mcp",{"commitSha":279},{"parentExtensionId":606,"repoId":607},"k17brg5egdw1jbncj1j4wfv3fh86n639","kd74zv63fryf9prygtq7gf4es986n22y",[215,602,603,601,592],{"evaluatedAt":610,"extractAt":611,"updatedAt":610},1778699724286,1778699234184,{"_creationTime":613,"_id":614,"community":615,"display":616,"identity":620,"providers":623,"relations":632,"tags":633,"workflow":634},1778699234184.6157,"k177tdbfgqmwhtaqv771f2ych586nne9",{"reviewCount":8},{"description":617,"installMethods":618,"name":619,"sourceUrl":593},"Worktree-first dev environment manager for issues, PRs, and features with optional tmux sessions",{"claudeCode":591},"Project Session Manager",{"basePath":621,"githubOwner":596,"githubRepo":597,"locale":18,"slug":622,"type":247},"skills/project-session-manager","project-session-manager",{"evaluate":624,"extract":631},{"promptVersionExtension":205,"promptVersionScoring":206,"score":511,"tags":625,"targetMarket":217,"tier":218},[626,627,628,629,277,215,630],"git","development-environment","workflow","tmux","developer-tool",{"commitSha":279,"license":280},{"parentExtensionId":606,"repoId":607},[277,215,630,627,626,629,628],{"evaluatedAt":635,"extractAt":611,"updatedAt":635},1778699613343,{"_creationTime":637,"_id":638,"community":639,"display":640,"identity":644,"providers":646,"relations":651,"tags":652,"workflow":653},1778699234184.6143,"k17cnx0m6a27fw52yvt4zsbsxh86nd1c",{"reviewCount":8},{"description":641,"installMethods":642,"name":643,"sourceUrl":593},"Configure popular MCP servers for enhanced agent capabilities",{"claudeCode":591},"mcp-setup",{"basePath":645,"githubOwner":596,"githubRepo":597,"locale":18,"slug":643,"type":247},"skills/mcp-setup",{"evaluate":647,"extract":650},{"promptVersionExtension":205,"promptVersionScoring":206,"score":511,"tags":648,"targetMarket":217,"tier":218},[603,602,215,294,649],"tooling",{"commitSha":279},{"parentExtensionId":606,"repoId":607},[294,215,602,603,649],{"evaluatedAt":654,"extractAt":611,"updatedAt":654},1778699492025]