This skill enables automated LLM evaluation, regression and security testing with Promptfoo, integrating into CI/CD to improve prompt quality and safety.
npx playbooks add skill phrazzld/claude-config --skill llm-evaluation
This skill enables automated LLM evaluation, regression and security testing with Promptfoo, integrating into CI/CD to improve prompt quality and safety.. This skill provides a specialized system prompt that configures your AI coding agent as a llm evaluation expert, with detailed methodology and structured output formats.
Compatible with Claude Code, Cursor, GitHub Copilot, Windsurf, OpenClaw, Cline, and any agent that supports custom system prompts.
This skill enables automated LLM evaluation, regression and security testing with Promptfoo, integrating into CI/CD to improve prompt quality and safety.