
A library for benchmarking, developing and deploying deep learning anomaly detection algorithms
Key Features • Docs • Notebooks • License
🌟 Announcing v2.2.0 Release! 🌟
We’re thrilled to announce the release of Anomalib v2.2.0, packed with new datasets, metrics, and performance improvements! Some of the highlights are: New datasets
- 3D-ADAM : A comprehensive dataset for 3D anomaly detection in additive manufacturing.
- BMAD : Benchmarks for Medical Anomaly Detection, featuring six datasets across five medical domains
New metrics
- PGn and PBn (CVPR2025) : Presorted good/bad metrics for more insightful performance evaluation.
- Histogram visualization of anomaly scores for better interpretability.
Other Improvements
- Faster coreset selection for PatchCore model, resulting in ~30% quicker training.
- Reduced memory usage for memory bank–based models like PatchCore, PaDiM, and DfKDE.
- Many more code and documentation updates.
We value your input! Please share feedback via GitHub Issues or our Discussions
Anomalib is a deep learning library that aims to collect state-of-the-art anomaly detection algorithms for benchmarking on both public and private datasets. Anomalib provides several ready-to-use implementations of anomaly detection algorithms described in the recent literature, as well as a set of tools that facilitate the development and implementation of custom models. The library has a strong focus on visual anomaly detection, where the goal of the algorithm is to detect and/or localize anomalies within images or videos in a dataset. Anomalib is constantly updated with new algorithms and training/inference extensions, so keep checking!
Anomalib can be installed from PyPI. We recommend using a virtual environment and a modern package installer like uv or pip.
For a standard installation, you can use uv or pip. This will install the latest version of Anomalib with its core dependencies. PyTorch will be installed based on its default behavior, which usually works for CPU and standard CUDA setups.
# With uv
uv pip install anomalib
# Or with pip
pip install anomalib
For more control over the installation, such as specifying the PyTorch backend (e.g., XPU, CUDA and ROCm) or installing extra dependencies for specific models, see the advanced options below.
# CPU support (default, works on all platforms)
uv pip install "anomalib[cpu]"
# CUDA 12.4 support (Linux/Windows with NVIDIA GPU)
uv pip install "anomalib[cu124]"
# CUDA 12.1 support (Linux/Windows with NVIDIA GPU)
uv pip install "anomalib[cu121]"
# CUDA 11.8 support (Linux/Windows with NVIDIA GPU)
uv pip install "anomalib[cu118]"
# ROCm support (Linux with AMD GPU)
uv pip install "anomalib[rocm]"
# Intel XPU support (Linux with Intel GPU)
uv pip install "anomalib[xpu]"
**Using `pip`:**
The same extras can be used with `pip`:
pip install "anomalib[cu124]"
# Example: Install with OpenVINO support and CUDA 12.4
uv pip install "anomalib[openvino,cu124]"
# Example: Install all optional dependencies for a CPU-only setup
uv pip install "anomalib[full,cpu]"
Here is a list of available optional dependency groups:
| Extra | Description | Purpose |
| :------------ | :--------------------------------------- | :------------------------------------------ |
| `[openvino]` | Intel OpenVINO optimization | For accelerated inference on Intel hardware |
| `[clip]` | Vision-language models | `winclip` |
| `[vlm]` | Vision-language model backends | Advanced VLM features |
| `[loggers]` | Experiment tracking (wandb, comet, etc.) | For experiment management |
| `[notebooks]` | Jupyter notebook support | For running example notebooks |
| `[full]` | All optional dependencies | All optional features |
git clone https://github.com/open-edge-platform/anomalib.git
cd anomalib
# Create the virtual environment
uv venv
# Sync with the lockfile for a specific backend (e.g., CPU)
uv sync --extra cpu
# Or for a different backend like CUDA 12.4
uv sync --extra cu124
# To set up a full development environment
uv sync --extra dev --extra cpu
**Using `pip`:**
git clone https://github.com/open-edge-platform/anomalib.git
cd anomalib
# Install in editable mode with a specific backend
pip install -e ".[cpu]"
# Install with development dependencies
pip install -e ".[dev,cpu]"
Anomalib supports both API and CLI-based training approaches:
from anomalib.data import MVTecAD
from anomalib.models import Patchcore
from anomalib.engine import Engine
# Initialize components
datamodule = MVTecAD()
model = Patchcore()
engine = Engine()
# Train the model
engine.fit(datamodule=datamodule, model=model)
# Train with default settings
anomalib train --model Patchcore --data anomalib.data.MVTecAD
# Train with custom category
anomalib train --model Patchcore --data anomalib.data.MVTecAD --data.category transistor
# Train with config file
anomalib train --config path/to/config.yaml
Anomalib provides multiple inference options including Torch, Lightning, Gradio, and OpenVINO. Here's how to get started:
# Load model and make predictions
predictions = engine.predict(
datamodule=datamodule,
model=model,
ckpt_path="path/to/checkpoint.ckpt",
)
# Basic prediction
anomalib predict --model anomalib.models.Patchcore \
--data anomalib.data.MVTecAD \
--ckpt_path path/to/model.ckpt
# Prediction with results
anomalib predict --model anomalib.models.Patchcore \
--data anomalib.data.MVTecAD \
--ckpt_path path/to/model.ckpt \
--return_predictions
📘 Note: For advanced inference options including Gradio and OpenVINO, check our Inference Documentation.
[!Note] Currently, only single GPU training is supported on Intel GPUs. These commands were tested on Arc 750 and Arc 770.
Ensure that you have PyTorch with XPU support installed. For more information, please refer to the PyTorch XPU documentation
from anomalib.data import MVTecAD
from anomalib.engine import Engine, SingleXPUStrategy, XPUAccelerator
from anomalib.models import Stfpm
engine = Engine(
strategy=SingleXPUStrategy(),
accelerator=XPUAccelerator(),
)
engine.train(Stfpm(), datamodule=MVTecAD())
anomalib train --model Padim --data MVTecAD --trainer.accelerator xpu --trainer.strategy xpu_single
Anomalib supports hyperparameter optimization (HPO) using Weights & Biases and Comet.ml.
# Run HPO with Weights & Biases
anomalib hpo --backend WANDB --sweep_config tools/hpo/configs/wandb.yaml
📘 Note: For detailed HPO configuration, check our HPO Documentation.
Track your experiments with popular logging platforms through PyTorch Lightning loggers:
Enable logging in your config file to track:
📘 Note: For logging setup, see our Logging Documentation.
Evaluate and compare model performance across different datasets:
# Run benchmarking with default configuration
anomalib benchmark --config tools/experimental/benchmarking/sample.yaml
💡 Tip: Check individual model performance in their respective README files:
If you find Anomalib useful in your research or work, please cite:
@inproceedings{akcay2022anomalib,
title={Anomalib: A deep learning library for anomaly detection},
author={Akcay, Samet and Ameln, Dick and Vaidya, Ashwin and Lakshmanan, Barath and Ahuja, Nilesh and Genc, Utku},
booktitle={2022 IEEE International Conference on Image Processing (ICIP)},
pages={1706--1710},
year={2022},
organization={IEEE}
}
We welcome contributions! Check out our Contributing Guide to get started.
Thank you to all our contributors!