Scholar Evaluation
技能 已验证 活跃Systematically evaluate scholarly work using the ScholarEval framework, providing structured assessment across research quality dimensions including problem formulation, methodology, analysis, and writing with quantitative scoring and actionable feedback.
To provide a structured, framework-based method for evaluating the quality and rigor of academic papers, research proposals, and other scholarly writing.
功能
- Systematic evaluation using ScholarEval framework
- Assessment across quality dimensions (problem, methodology, writing, etc.)
- Provides quantitative scoring and qualitative feedback
- Supports various scholarly work types (papers, proposals, reviews)
使用场景
- Evaluating research papers for publication readiness
- Assessing research proposals for funding applications
- Reviewing literature reviews for comprehensiveness and quality
- Providing structured feedback on academic writing
非目标
- Replacing domain-specific expertise
- Conducting the primary research itself
- Evaluating non-scholarly content
安装
npx skills add K-Dense-AI/claude-scientific-skills通过 npx 运行 Vercel skills CLI(skills.sh)— 需要本地安装 Node.js,以及至少一个兼容 skills 的智能体(Claude Code、Cursor、Codex 等)。前提是仓库遵循 agentskills.io 格式。
质量评分
已验证类似扩展
Literature Review
100Conduct comprehensive, systematic literature reviews using multiple academic databases (PubMed, arXiv, bioRxiv, Semantic Scholar, etc.). This skill should be used when conducting systematic literature reviews, meta-analyses, research synthesis, or comprehensive literature searches across biomedical, scientific, and technical domains. Creates professionally formatted markdown documents and PDFs with verified citations in multiple citation styles (APA, Nature, Vancouver, etc.).
Notion 研究文档
100跨 Notion 工作区进行搜索,综合多个页面的发现,并创建保存为新 Notion 页面的全面研究文档。将分散的信息转化为具有正确引用和可操作见解的结构化报告。
Survey Theoretical Literature
99Survey and synthesize theoretical literature on a specific topic, identifying seminal papers, key results, open problems, and cross-domain connections. Use when starting research on an unfamiliar theoretical topic, writing a literature review for a paper or thesis, identifying open problems and research gaps, finding cross-domain connections, or evaluating the novelty of a proposed theoretical contribution against existing work.
Evaluating Llms Harness
99Evaluates LLMs across 60+ academic benchmarks (MMLU, HumanEval, GSM8K, TruthfulQA, HellaSwag). Use when benchmarking model quality, comparing models, reporting academic results, or tracking training progress. Industry standard used by EleutherAI, HuggingFace, and major labs. Supports HuggingFace, vLLM, APIs.
Lm Evaluation Harness
98Evaluates LLMs across 60+ academic benchmarks (MMLU, HumanEval, GSM8K, TruthfulQA, HellaSwag). Use when benchmarking model quality, comparing models, reporting academic results, or tracking training progress. Industry standard used by EleutherAI, HuggingFace, and major labs. Supports HuggingFace, vLLM, APIs.
Alterlab Openalex
98Query and analyze scholarly literature using the OpenAlex database. This skill should be used when searching for academic papers, analyzing research trends, finding works by authors or institutions, tracking citations, discovering open access publications, or conducting bibliometric analysis across 240M+ scholarly works. Use for literature searches, research output analysis, citation analysis, and academic database queries. Part of the AlterLab Academic Skills suite.