Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
16 changes: 16 additions & 0 deletions skills/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -12,6 +12,22 @@ applications using eval results.

## Installation

### For Claude.ai (Web)

1. Download the `skills/deepeval` folder from this repository.
2. Zip the folder.
3. In Claude.ai, navigate to **Settings > Capabilities > Skills**.
4. Click **Upload skill** and select your zipped folder.

### For Claude Code (Local CLI)

Download or clone the `skills/deepeval` folder inside the skills folder and place it directly into your local project's skills directory:

```bash
mkdir -p .claude/skills/
cp -r path/to/downloaded/deepeval .claude/skills/
```

### Cursor Plugin

This repository includes a Cursor plugin manifest that points to `./skills/`.
Expand Down
25 changes: 0 additions & 25 deletions skills/deepeval/README.md

This file was deleted.

10 changes: 10 additions & 0 deletions skills/deepeval/SKILL.md
Original file line number Diff line number Diff line change
Expand Up @@ -27,6 +27,16 @@ Use this skill to add an end-to-end eval loop to AI applications:
instrument the app, generate or reuse a dataset, create a committed pytest eval
suite, run evals, and iterate on failures.

## Workflow Summary

1. Inspect the target app and existing DeepEval usage.
2. Ask the required intake questions.
3. Reuse existing metrics and datasets when available.
4. Generate or import goldens.
5. Add minimal tracing and a pytest eval suite.
6. Run `deepeval test run`.
7. Iterate for the requested number of rounds, defaulting to 5.

## Core Principles

1. Prefer the smallest committed pytest eval suite that the user can rerun
Expand Down
Loading