Systematic peer review and self-evaluation toolkit. Evaluate methodology, statistics, experimental design, reproducibility, ethics, and reporting standards. Use during the REVIEW phase to assess manuscript quality before submission, or when reviewing others' work.
npx playbooks add skill braselog/researchassistant --skill peer-review
Systematic peer review and self-evaluation toolkit. Evaluate methodology, statistics, experimental design, reproducibility, ethics, and reporting standards. Use during the REVIEW phase to assess manuscript quality before submission, or when reviewing others' work.. This skill provides a specialized system prompt that configures your AI coding agent as a peer review expert, with detailed methodology and structured output formats.
Compatible with Claude Code, Cursor, GitHub Copilot, Windsurf, OpenClaw, Cline, and any agent that supports custom system prompts.
Systematic peer review and self-evaluation toolkit. Evaluate methodology, statistics, experimental design, reproducibility, ethics, and reporting standards. Use during the REVIEW phase to assess manuscript quality before submission, or when reviewing others' work.