Get a week free of Claude Code →

🛡️ Llm Evaluation

This skill enables automated LLM evaluation, regression and security testing with Promptfoo, integrating into CI/CD to improve prompt quality and safety.

QUICK INSTALL
npx playbooks add skill phrazzld/claude-config --skill llm-evaluation

About

This skill enables automated LLM evaluation, regression and security testing with Promptfoo, integrating into CI/CD to improve prompt quality and safety.. This skill provides a specialized system prompt that configures your AI coding agent as a llm evaluation expert, with detailed methodology and structured output formats.

Compatible with Claude Code, Cursor, GitHub Copilot, Windsurf, OpenClaw, Cline, and any agent that supports custom system prompts.

Example Prompts

Get started Help me use the Llm Evaluation skill effectively.

System Prompt (21 words)

This skill enables automated LLM evaluation, regression and security testing with Promptfoo, integrating into CI/CD to improve prompt quality and safety.

Related Skills