Promptfoo Evaluation
skill from daymade/claude-code-skills
About
--- name: promptfoo-evaluation description: Configures and runs LLM evaluation using Promptfoo framework. Use when setting up prompt testing, creating evaluation configs (promptfooconfig.yaml), writing Python custom assertions, implementing llm-rubric for LLM-as-judge, or managing few-shot examples in prompts. Triggers on keywords like "promptfoo", "eval", "LLM evaluation", "prompt testing", or "model comparison". --- # Promptfoo Evaluation ## Overview This skill provides guidance for configuring and running LLM evaluations using [Promptfoo](https://www.promptfoo.dev/), an open-source CLI tool for testing and comparing LLM outputs. ## Quick Start ```bash # Initialize a new evaluation project npx promptfoo@latest init # Run evaluation npx promptfoo@latest eval # View results in browser npx promptfoo@latest view ``` ## Configuration Structure A typical Promptfoo project structure: ``` project/ ├── promptfooconfig.yaml # Main configuration ├── prompts/ │ ├── system.md
Quick Start
Manual Installation
No automatic installation available. Please visit the source repository for installation instructions.
View Installation Instructions