This skill guides configuring and running LLM evaluations with Promptfoo, enabling prompt testing, rubric-based judging, and custom assertions.
npx playbooks add skill daymade/claude-code-skills --skill promptfoo-evaluation
This skill guides configuring and running LLM evaluations with Promptfoo, enabling prompt testing, rubric-based judging, and custom assertions.. This skill provides a specialized system prompt that configures your AI coding agent as a promptfoo evaluation expert, with detailed methodology and structured output formats.
Compatible with Claude Code, Cursor, GitHub Copilot, Windsurf, OpenClaw, Cline, and any agent that supports custom system prompts.
This skill guides configuring and running LLM evaluations with Promptfoo, enabling prompt testing, rubric-based judging, and custom assertions.
Weekly roundup of top Claude Code skills, MCP servers, and AI coding tips.