deepeval-testing

Skill for pytest-style LLM evaluation with DeepEval.

Safety Notice

This listing is imported from skills.sh public index metadata. Review upstream SKILL.md and repository scripts before running.

Copy this and send it to your AI assistant to learn

Install skill "deepeval-testing" with this command: npx skills add vanman2024/ai-dev-marketplace/vanman2024-ai-dev-marketplace-deepeval-testing

DeepEval Testing

Skill for pytest-style LLM evaluation with DeepEval.

Overview

DeepEval provides:

  • pytest-compatible LLM tests

  • Built-in metrics (faithfulness, relevance, toxicity)

  • Custom metric creation

  • Async test execution

Use When

This skill is automatically invoked when:

  • Creating pytest-style LLM tests

  • Evaluating RAG quality

  • Measuring faithfulness/relevance

  • Building custom metrics

Available Scripts

Script Description

scripts/init-deepeval.sh

Initialize DeepEval project

scripts/run-tests.sh

Run DeepEval tests

Available Templates

Template Description

templates/conftest.py

pytest configuration

templates/test_basic.py

Basic test structure

templates/test_rag.py

RAG evaluation tests

Built-in Metrics

  • AnswerRelevancyMetric

  • FaithfulnessMetric

  • ContextualRelevancyMetric

  • HallucinationMetric

  • ToxicityMetric

  • BiasMetric

Source Transparency

This detail page is rendered from real SKILL.md content. Trust labels are metadata-based hints, not a safety guarantee.

Related Skills

Related by shared tags or category signals.

Coding

document-parsers

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

stt-integration

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

model-routing-patterns

No summary provided by upstream source.

Repository SourceNeeds Review
Coding

react-email-templates

No summary provided by upstream source.

Repository SourceNeeds Review