DeepEval
Verified Open SourceOpen-source LLM evaluation framework for CI/CD pipelines
DeepEval is an open-source evaluation framework for testing and monitoring LLM applications. It provides 14+ evaluation metrics, integrates with pytest, and enables continuous LLM quality testing in CI/CD workflows.
Product Overview
Use Cases
- LLM Unit Testing
- CI/CD for AI
- RAG Quality Testing
- Hallucination Detection
Ideal For
AI EngineersQA TeamsMLOps Engineers
Architecture Fit
Enterprise ReadySelf HostedCloud NativeAPI FirstMulti-Agent CompatibleKubernetes SupportOpen Source
Technical Details
- Deployment Model
- self-hosted
Screenshots
No screenshots available yet.
Community Feedback
Loadingβ¦
Login to leave feedback on this product.