LogoLogo
HomeBlogGitHub
latest
latest
  • New DOCS
  • What is Evidently?
  • Get Started
    • Evidently Cloud
      • Quickstart - LLM tracing
      • Quickstart - LLM evaluations
      • Quickstart - Data and ML checks
      • Quickstart - No-code evaluations
    • Evidently OSS
      • OSS Quickstart - LLM evals
      • OSS Quickstart - Data and ML monitoring
  • Presets
    • All Presets
    • Data Drift
    • Data Quality
    • Target Drift
    • Regression Performance
    • Classification Performance
    • NoTargetPerformance
    • Text Evals
    • Recommender System
  • Tutorials and Examples
    • All Tutorials
    • Tutorial - Tracing
    • Tutorial - Reports and Tests
    • Tutorial - Data & ML Monitoring
    • Tutorial - LLM Evaluation
    • Self-host ML Monitoring
    • LLM as a judge
    • LLM Regression Testing
  • Setup
    • Installation
    • Evidently Cloud
    • Self-hosting
  • User Guide
    • 📂Projects
      • Projects overview
      • Manage Projects
    • 📶Tracing
      • Tracing overview
      • Set up tracing
    • 🔢Input data
      • Input data overview
      • Column mapping
      • Data for Classification
      • Data for Recommendations
      • Load data to pandas
    • 🚦Tests and Reports
      • Reports and Tests Overview
      • Get a Report
      • Run a Test Suite
      • Evaluate Text Data
      • Output formats
      • Generate multiple Tests or Metrics
      • Run Evidently on Spark
    • 📊Evaluations
      • Evaluations overview
      • Generate snapshots
      • Run no code evals
    • 🔎Monitoring
      • Monitoring overview
      • Batch monitoring
      • Collector service
      • Scheduled evaluations
      • Send alerts
    • 📈Dashboard
      • Dashboard overview
      • Pre-built Tabs
      • Panel types
      • Adding Panels
    • 📚Datasets
      • Datasets overview
      • Work with Datasets
    • 🛠️Customization
      • Data drift parameters
      • Embeddings drift parameters
      • Feature importance in data drift
      • Text evals with LLM-as-judge
      • Text evals with HuggingFace
      • Add a custom text descriptor
      • Add a custom drift method
      • Add a custom Metric or Test
      • Customize JSON output
      • Show raw data in Reports
      • Add text comments to Reports
      • Change color schema
    • How-to guides
  • Reference
    • All tests
    • All metrics
      • Ranking metrics
    • Data drift algorithm
    • API Reference
      • evidently.calculations
        • evidently.calculations.stattests
      • evidently.metrics
        • evidently.metrics.classification_performance
        • evidently.metrics.data_drift
        • evidently.metrics.data_integrity
        • evidently.metrics.data_quality
        • evidently.metrics.regression_performance
      • evidently.metric_preset
      • evidently.options
      • evidently.pipeline
      • evidently.renderers
      • evidently.report
      • evidently.suite
      • evidently.test_preset
      • evidently.test_suite
      • evidently.tests
      • evidently.utils
  • Integrations
    • Integrations
      • Evidently integrations
      • Notebook environments
      • Evidently and Airflow
      • Evidently and MLflow
      • Evidently and DVCLive
      • Evidently and Metaflow
  • SUPPORT
    • Migration
    • Contact
    • F.A.Q.
    • Telemetry
    • Changelog
  • GitHub Page
  • Website
Powered by GitBook
On this page
  • Metric Presets
  • Test Presets
  • Individual Metrics and Tests
  1. Presets

All Presets

An overview of the evaluations you can do with Evidently.

PreviousPresetsNextData Drift

Last updated 2 months ago

You are looking at the old Evidently documentation. Check the newer version .

Evidently has several pre-built reports and test suites. We call them Presets. Each preset evaluates or tests a particular aspect of the data or model quality.

This page links to the description of each preset. To see the code and interactive examples, head to instead.

Metric Presets

Metric presets are pre-built reports that help with visual exploration, debugging and documentation of the data and model performance. You can also use them to calculate and log metrics as JSON or Python dictionary.

Shows the dataset statistics and feature behavior. Requirements: model inputs.

Explores the distribution shift in the model features. Requirements: model inputs, a reference dataset.

Explores the distribution shift in the model predictions. Requirements: model predictions and/or target, a reference dataset.

Evaluates the classification model quality and errors. Requirements: model predictions and true labels.

Evaluates the regression model quality and errors. Requirements: model predictions and actuals.

Evaluates text data drift and descriptive statistics. Requirements: model inputs (raw text data)

Test Presets

Test presets are pre-built test suites that perform structured data and model checks as part of the pipeline.

Tests the model performance without ground truth or actuals. Requirements: model inputs, predictions, a reference dataset.

Tests if the data quality is suitable for (re)training. Checks nulls, duplicates, etc. Requirements: model inputs.

Tests if a data batch is similar to reference. Checks schema, data ranges, etc. Requirements: inputs, a reference dataset.

Tests for distribution drift per column and overall dataset drift. Requirements: model inputs, a reference dataset.

Tests the performance of the regression model against expectation. Requirements: model predictions and actuals.

Tests the performance of a multi-class classification model against expectation. Requirements: model predictions, true labels.

Tests the performance of a binary classification model against expectation. Requirements: model predictions, true labels.

Tests the performance of a binary classification model at top-K. Requirements: model predictions, true labels.

Individual Metrics and Tests

You can also create custom test suites and reports from individual metrics and tests. You can explore 100+ and .

here
example notebooks
available tests
metrics
Data Quality
Data Drift
Target Drift
Classification
Regression
Text Overview
NoTargetPerformance
Data Quality
Data Stability
Multi-class Classification
Binary Classification
Binary Classification top-K
Data Drift
Regression