This repository provides precise, efficient, and extensible implementations of the popular metrics for generative model evaluation, including:
- Inception Score (ISC)
- Frรฉchet Inception Distance (FID)
- Kernel Inception Distance (KID)
- Precision and Recall (PRC)
- Perceptual Path Length (PPL)
Numerical Precision: Unlike many other reimplementations, the values produced by torch-fidelity match reference implementations up to floating point's machine precision. This allows using torch-fidelity for reporting metrics in papers instead of scattered and slow reference implementations. Read more about numerical precision
Efficiency: Feature sharing between different metrics saves recomputation time, and an additional caching level avoids recomputing features and statistics whenever possible. High efficiency allows using torch-fidelity in the training loop, for example at the end of every epoch. Read more about efficiency
Extensibility: Going beyond 2D image generation is easy due to high modularity and abstraction of the metrics from input data, models, and feature extractors. For example, one can swap out InceptionV3 feature extractor for a one accepting 3D scan volumes, such as used in MRI. Read more about extensibility
TLDR; fast and reliable GAN evaluation in PyTorch
Installation
pip install torch-fidelity
See also: Installing the latest GitHub code
Usage Examples with Command Line
Below are three examples of using torch-fidelity to evaluate metrics from the command line. See more examples in the documentation.
Simple
Inception Score of CIFAR-10 training split:
> fidelity --gpu 0 --isc --input1 cifar10-train
inception_score_mean: 11.23678
inception_score_std: 0.09514061
Medium
Inception Score of a directory of images stored in ~/images/
:
> fidelity --gpu 0 --isc --input1 ~/images/
Pro
Efficient computation of ISC and PPL for input1
, and FID, KID, PRC between a generative model stored in ~/generator.onnx
and CIFAR-10 training split:
> fidelity \
--gpu 0 \
--isc \
--fid \
--kid \
--ppl \
--prc \
--input1 ~/generator.onnx \
--input1-model-z-type normal \
--input1-model-z-size 128 \
--input1-model-num-samples 50000 \
--input2 cifar10-train
See also: Other usage examples
Quick Start with Python API
When it comes to tracking the performance of generative models as they train, evaluating metrics after every epoch
becomes prohibitively expensive due to long computation times.
torch_fidelity
tackles this problem by making full use
of caching to avoid recomputing common features and per-metric statistics whenever possible.
Computing all metrics for 50000 32x32 generated images and cifar10-train
takes only 2 min 26 seconds on NVIDIA P100
GPU, compared to >10 min if using original codebases.
Thus, computing metrics 20 times over the whole training cycle makes overall training time just one hour longer.
In the following example, assume unconditional image generation setting with CIFAR-10, and the generative model
generator
, which takes a 128-dimensional standard normal noise vector.
First, import the module:
import torch_fidelity
Add the following lines at the end of epoch evaluation:
wrapped_generator = torch_fidelity.GenerativeModelModuleWrapper(generator, 128, 'normal', 0)
metrics_dict = torch_fidelity.calculate_metrics(
input1=wrapped_generator,
input2='cifar10-train',
cuda=True,
isc=True,
fid=True,
kid=True,
prc=True,
verbose=False,
)
The resulting dictionary with computed metrics can logged directly to tensorboard, wandb, or console:
print(metrics_dict)
Output:
{
'inception_score_mean': 11.23678,
'inception_score_std': 0.09514061,
'frechet_inception_distance': 18.12198,
'kernel_inception_distance_mean': 0.01369556,
'kernel_inception_distance_std': 0.001310059
'precision': 0.51369556,
'recall': 0.501310059
}
See also: Full API reference
Example of Integration with the Training Loop
Refer to sngan_cifar10.py for a complete training example.
Evolution of fixed generator latents in the example:
A generator checkpoint resulting from training the example can be downloaded here.
Troubleshooting
WARNING: The script fidelity is installed in '' which is not on PATH.
Suggests that the standalone fidelity
tool will not be available unless the above path is added to the PATH
environment variable. If modifying it is undesirable, the tool can still be called by its full path: <SOMEPATH>/fidelity
.
Citation
Citation is recommended to reinforce the evaluation protocol in works relying on torch-fidelity. To ensure reproducibility when citing this repository, use the following BibTeX:
@misc{obukhov2020torchfidelity,
author={Anton Obukhov and Maximilian Seitzer and Po-Wei Wu and Semen Zhydenko and Jonathan Kyl and Elvis Yu-Jing Lin},
year=2020,
title={High-fidelity performance metrics for generative models in PyTorch},
url={https://github.com/toshas/torch-fidelity},
publisher={Zenodo},
version={v0.3.0},
doi={10.5281/zenodo.4957738},
note={Version: 0.3.0, DOI: 10.5281/zenodo.4957738}
}