jpskill.com
💼 ビジネス コミュニティ 🟡 少し慣れが必要 👤 経営者・事業責任者・マーケ

💼 Scholar Evaluation

scholar-evaluation

論文や提案書、文献レビューなどの学

⏱ 営業メール作成 15分/通 → 1分/通

📺 まず動画で見る(YouTube)

▶ 【自動化】AIガチ勢の最新活用術6選がこれ1本で丸分かり!【ClaudeCode・AIエージェント・AI経営・Skills・MCP】 ↗

※ jpskill.com 編集部が参考用に選んだ動画です。動画の内容と Skill の挙動は厳密には一致しないことがあります。

📜 元の英語説明(参考)

Structured scholarly-work evaluation for papers, proposals, literature reviews, methods sections, evidence quality, citation support, and research-writing feedback.

🇯🇵 日本人クリエイター向け解説

一言でいうと

論文や提案書、文献レビューなどの学

※ jpskill.com 編集部が日本のビジネス現場向けに補足した解説です。Skill本体の挙動とは独立した参考情報です。

⚠️ ダウンロード・利用は自己責任でお願いします。当サイトは内容・動作・安全性について責任を負いません。

🎯 このSkillでできること

下記の説明文を読むと、このSkillがあなたに何をしてくれるかが分かります。Claudeにこの分野の依頼をすると、自動で発動します。

📦 インストール方法 (3ステップ)

  1. 1. 上の「ダウンロード」ボタンを押して .skill ファイルを取得
  2. 2. ファイル名の拡張子を .skill から .zip に変えて展開(macは自動展開可)
  3. 3. 展開してできたフォルダを、ホームフォルダの .claude/skills/ に置く
    • · macOS / Linux: ~/.claude/skills/
    • · Windows: %USERPROFILE%\.claude\skills\

Claude Code を再起動すれば完了。「このSkillを使って…」と話しかけなくても、関連する依頼で自動的に呼び出されます。

詳しい使い方ガイドを見る →
最終更新
2026-05-17
取得日時
2026-05-17
同梱ファイル
1

💬 こう話しかけるだけ — サンプルプロンプト

  • Scholar Evaluation で、私のビジネスを分析して改善案を3つ提案して
  • Scholar Evaluation を使って、来週の会議用の資料を作って
  • Scholar Evaluation で、現状の課題を整理してアクションプランに落として

これをClaude Code に貼るだけで、このSkillが自動発動します。

📖 Claude が読む原文 SKILL.md(中身を展開)

この本文は AI(Claude)が読むための原文(英語または中国語)です。日本語訳は順次追加中。

Scholar Evaluation

Use this skill to evaluate academic or scientific work with a repeatable rubric.

When to Use

  • Reviewing a research paper, proposal, thesis chapter, or literature review.
  • Checking whether claims are supported by cited evidence.
  • Evaluating methodology, study design, analysis, or limitations.
  • Comparing two or more papers for quality or relevance.
  • Producing structured feedback for revision.

Evaluation Scope

Start by identifying the artifact:

  • empirical research paper
  • theoretical paper
  • technical report
  • systematic or narrative literature review
  • research proposal
  • thesis or dissertation chapter
  • conference abstract or short paper

Then choose scope:

  • comprehensive: all rubric dimensions
  • targeted: one or two dimensions, such as method or citations
  • comparative: rank multiple works against the same rubric

Rubric

Score each applicable dimension from 1 to 5:

  • 5: excellent; clear, rigorous, and publication-ready
  • 4: good; minor improvements needed
  • 3: adequate; meaningful gaps but usable
  • 2: weak; substantial revision needed
  • 1: poor; major validity or clarity problems

Use N/A for dimensions that do not apply.

1. Problem and Research Question

  • Is the problem clear and specific?
  • Is the contribution meaningful?
  • Are scope and assumptions explicit?
  • Does the question match the claimed contribution?

2. Literature and Context

  • Is relevant prior work covered?
  • Does the work synthesize rather than merely list sources?
  • Are gaps accurately identified?
  • Are recent and foundational sources balanced?

3. Methodology

  • Does the method answer the research question?
  • Are design choices justified?
  • Are variables, datasets, participants, or materials described clearly?
  • Could another researcher reproduce the work?
  • Are ethical and practical constraints acknowledged?

4. Data and Evidence

  • Are data sources credible and appropriate?
  • Is sample size or corpus coverage adequate?
  • Are inclusion, exclusion, and preprocessing decisions documented?
  • Are missing data and bias risks discussed?

5. Analysis

  • Are statistical, qualitative, or computational methods appropriate?
  • Are baselines and controls fair?
  • Are uncertainty, sensitivity, or robustness checks included when needed?
  • Are alternative explanations considered?

6. Results and Interpretation

  • Are results clearly presented?
  • Do claims stay within the evidence?
  • Are figures, tables, and metrics understandable?
  • Are negative or null results handled honestly?

7. Limitations and Threats to Validity

  • Are limitations specific rather than generic?
  • Are internal, external, construct, and conclusion-validity risks addressed?
  • Does the paper distinguish speculation from demonstrated results?

8. Writing and Structure

  • Is the argument easy to follow?
  • Are sections organized around the research question?
  • Are definitions and notation clear?
  • Is the tone precise and scholarly?

9. Citations

  • Do cited papers support the claims attached to them?
  • Are primary sources used where possible?
  • Are reviews labeled as reviews?
  • Are preprints labeled as preprints?
  • Are citation metadata and links correct?

Review Process

  1. Read the abstract, introduction, figures, and conclusion for claimed contribution.
  2. Read methods and results for evidence quality.
  3. Check the strongest claims against cited sources.
  4. Score each applicable dimension.
  5. Separate critical blockers from revision suggestions.
  6. End with concrete next edits.

Output Template

# Scholar Evaluation: <Artifact>

## Overall Assessment

- Overall score: <1-5 or N/A>
- Confidence: <high | medium | low>
- Summary: <3-5 sentences>

## Dimension Scores

| Dimension | Score | Evidence | Revision priority |
| --- | ---: | --- | --- |
| Problem and question |  |  |  |
| Literature and context |  |  |  |
| Methodology |  |  |  |
| Data and evidence |  |  |  |
| Analysis |  |  |  |
| Results and interpretation |  |  |  |
| Limitations |  |  |  |
| Writing and structure |  |  |  |
| Citations |  |  |  |

## Critical Issues

## Recommended Revisions

## Evidence Checks Needed

Pitfalls

  • Do not use the score as a substitute for concrete feedback.
  • Do not penalize a paper for omitting a dimension outside its scope.
  • Do not treat citation count, venue, or author reputation as proof of quality.
  • Do not accept unsupported claims just because they appear in the abstract.