Hypogenic
技能 已验证 活跃Automated LLM-driven hypothesis generation and testing on tabular datasets. Use when you want to systematically explore hypotheses about patterns in empirical data (e.g., deception detection, content analysis). Combines literature insights with data-driven hypothesis testing. For manual hypothesis formulation use hypothesis-generation; for creative ideation use scientific-brainstorming.
To accelerate scientific discovery by automating the systematic exploration and testing of hypotheses derived from empirical data and existing literature.
功能
- Automated hypothesis generation from tabular data
- Integration of literature insights for hypothesis refinement
- Support for multiple hypothesis generation methods (data-driven, literature-integrated, union)
- Flexible configuration via YAML and prompt templates
- CLI and Python API for programmatic control
使用场景
- Generating testable hypotheses for empirical research without prior theoretical frameworks
- Testing competing hypotheses by combining literature insights with data patterns
- Accelerating discovery in domains like deception detection, AI content identification, and predictive modeling
- Validating or extending existing theories with new empirical evidence
非目标
- Manual hypothesis formulation (use hypothesis-generation skill)
- Creative ideation without data grounding (use scientific-brainstorming skill)
- General-purpose data analysis without a hypothesis-driven approach
- Replacing human scientific intuition entirely; acts as an assistant
安装
npx skills add K-Dense-AI/claude-scientific-skills通过 npx 运行 Vercel skills CLI(skills.sh)— 需要本地安装 Node.js,以及至少一个兼容 skills 的智能体(Claude Code、Cursor、Codex 等)。前提是仓库遵循 agentskills.io 格式。
质量评分
已验证类似扩展
Arize Prompt Optimization
100Optimizes, improves, and debugs LLM prompts using production trace data, evaluations, and annotations. Extracts prompts from spans, gathers performance signal, and runs a data-driven optimization loop using the ax CLI. Use when the user mentions optimize prompt, improve prompt, make AI respond better, improve output quality, prompt engineering, prompt tuning, or system prompt improvement.
Gtars
99High-performance toolkit for genomic interval analysis in Rust with Python bindings. Use when working with genomic regions, BED files, coverage tracks, overlap detection, tokenization for ML models, or fragment analysis in computational genomics and machine learning applications.
Alterlab Hypothesis Gen
95Part of the AlterLab Academic Skills suite. Structured hypothesis formulation from observations. Use when you have experimental observations or data and need to formulate testable hypotheses with predictions, propose mechanisms, and design experiments to test them. Follows scientific method framework. For open-ended ideation use scientific-brainstorming; for automated LLM-driven hypothesis testing on datasets use hypogenic.
Survey Insect Population
100Design and execute insect population surveys covering survey design, sampling methods, field execution, specimen identification, diversity index calculation including Shannon-Wiener and Simpson indices, statistical analysis, and reporting. Covers defining survey objectives, selecting study sites, determining sampling intensity and replication, choosing sampling methods appropriate to target taxa, standardizing collection effort, recording environmental covariates, identifying specimens to the lowest practical taxonomic level, calculating species richness, Shannon-Wiener diversity (H'), Simpson diversity (1-D), evenness, rarefaction curves, multivariate ordination, and producing survey reports with species lists and conservation implications. Use when conducting baseline biodiversity assessments, monitoring insect populations over time, comparing insect communities across habitats or treatments, assessing environmental impact, or supporting conservation planning with quantitative ecological data.
Fit Drift Diffusion Model
100Fit cognitive drift-diffusion models (Ratcliff DDM) to reaction time and accuracy data with parameter estimation (drift rate, boundary separation, non-decision time), model comparison, and parameter recovery validation. Use when modeling binary decision-making with reaction time data, estimating cognitive parameters from experimental data, comparing sequential sampling model variants, or decomposing speed-accuracy tradeoff effects into latent cognitive components.
Measure Experiment Design
100Designs an A/B test or experiment with clear hypothesis, variants, success metrics, sample size, and duration. Use when planning experiments to validate product changes or test hypotheses.