Get a week free of Claude Code →

🧪 Promptfoo Evaluation

This skill guides configuring and running LLM evaluations with Promptfoo, enabling prompt testing, rubric-based judging, and custom assertions.

QUICK INSTALL
npx playbooks add skill daymade/claude-code-skills --skill promptfoo-evaluation

About

This skill guides configuring and running LLM evaluations with Promptfoo, enabling prompt testing, rubric-based judging, and custom assertions.. This skill provides a specialized system prompt that configures your AI coding agent as a promptfoo evaluation expert, with detailed methodology and structured output formats.

Compatible with Claude Code, Cursor, GitHub Copilot, Windsurf, OpenClaw, Cline, and any agent that supports custom system prompts.

Example Prompts

Get started Help me use the Promptfoo Evaluation skill effectively.

System Prompt (18 words)

This skill guides configuring and running LLM evaluations with Promptfoo, enabling prompt testing, rubric-based judging, and custom assertions.

Related Skills

Get the best new skills
in your inbox

Weekly roundup of top Claude Code skills, MCP servers, and AI coding tips.