Get a week free of Claude Code →

🔍 Peer Review

Systematic peer review and self-evaluation toolkit. Evaluate methodology, statistics, experimental design, reproducibility, ethics, and reporting standards. Use during the REVIEW phase to assess manuscript quality before submission, or when reviewing others' work.

QUICK INSTALL
npx playbooks add skill braselog/researchassistant --skill peer-review

About

Systematic peer review and self-evaluation toolkit. Evaluate methodology, statistics, experimental design, reproducibility, ethics, and reporting standards. Use during the REVIEW phase to assess manuscript quality before submission, or when reviewing others' work.. This skill provides a specialized system prompt that configures your AI coding agent as a peer review expert, with detailed methodology and structured output formats.

Compatible with Claude Code, Cursor, GitHub Copilot, Windsurf, OpenClaw, Cline, and any agent that supports custom system prompts.

Example Prompts

Get started Help me use the Peer Review skill effectively.

System Prompt (32 words)

Systematic peer review and self-evaluation toolkit. Evaluate methodology, statistics, experimental design, reproducibility, ethics, and reporting standards. Use during the REVIEW phase to assess manuscript quality before submission, or when reviewing others' work.

Related Skills