Back to skills
SkillHub ClubAnalyze Data & AITestingData / AIIntegration

deepeval

This skill provides comprehensive documentation for DeepEval, a pytest-based framework for evaluating LLM applications. It covers installation, core workflows for RAG and conversational AI testing, dataset management, and common patterns with practical code examples.

Packaged view

This page reorganizes the original catalog entry around fit, installability, and workflow context first. The original raw source lives below.

Stars
110
Hot score
94
Updated
March 19, 2026
Overall rating
A8.3
Composite score
5.8
Best-practice grade
B75.6

Install command

npx @skill-hub/cli install sammcj-agentic-coding-deepeval
llm-evaluationrag-testingpytest-integrationai-testingsynthetic-data

Repository

sammcj/agentic-coding

Skill path: Claude/skills/deepeval

This skill provides comprehensive documentation for DeepEval, a pytest-based framework for evaluating LLM applications. It covers installation, core workflows for RAG and conversational AI testing, dataset management, and common patterns with practical code examples.

Open repository

Best for

Primary workflow: Analyze Data & AI.

Technical facets: Testing, Data / AI, Integration.

Target audience: Testing teams looking for install-ready agent workflows..

License: Unknown.

Original source

Catalog source: SkillHub Club.

Repository owner: sammcj.

This is still a mirrored public skill entry. Review the repository before installing into production workflows.

What it helps with

  • Install deepeval into Claude Code, Codex CLI, Gemini CLI, or OpenCode workflows
  • Review https://github.com/sammcj/agentic-coding before adding deepeval to shared team environments
  • Use deepeval for testing workflows

Works across

Claude CodeCodex CLIGemini CLIOpenCode

Favorites: 0.

Sub-skills: 0.

Aggregator: No.

deepeval | SkillHub