IFQA: Interpretable Face Quality Assessment
Byungho Jo, Donghyeon Cho, In Kyu Park, Sungeun Hong
In WACV 2023
Paper:
Abstract: Existing face restoration models have relied on general assessment metrics that do not consider the characteristics of facial regions. Recent works have therefore assessed their methods using human studies, which is not scalable and involves significant effort. This paper proposes a novel face-centric metric based on an adversarial framework where a generator simulates face restoration and a discriminator assesses image quality. Specifically, our per-pixel discriminator enables interpretable evaluation that cannot be provided by traditional metrics. Moreover, our metric emphasizes facial primary regions considering that even minor changes to the eyes, nose, and mouth significantly affect human cognition. Our face-oriented metric consistently surpasses existing general or facial image quality assessment metrics by impressive margins. We demonstrate the generalizability of the proposed strategy in various architectural designs and challenging scenarios. Interestingly, we find that our IFQA can lead to performance improvement as an objective function.
- 64-bit Python 3.7 and PyTorch 1.7.0 (or later). See https://pytorch.org for PyTorch install instructions.
- Albumentations. See https://albumentations.ai/ for Albumentations install instructions.
IFQA is designed for evaluating the realness of faces. IFQA produces score maps of each pixel and we apply average to get final score.
You can produce quality scores using test.py
. For example:
# A single face image input.
python test.py --path=./faces0.png
# All images within a directory.
python test.py --path=./testA
@inproceedings{Jo2023,
author = {Byungho Jo and Donghyeon Cho and In Kyu Park and Sungeun Hong},
title = {IFQA: Interpretable Face Quality Assessment},
booktitle = {Proc. WACV},
year = {2023}
}
We thank Eunkyung Jo for helpful feedback on human study design and Jaejun Yoo for constructive comments on various experimental protocols.