The key insight of our approach is to reuse procedurally generated identities from a graphics pipeline and enhance their realism to reduce the domain gap. DigiFace1M provides an elaborate pipeline for generating synthetic identities and their variations, allowing us to obtain a large number of identities from this dataset. Additionally, we generate variations by interpolating between multiple images of an identity within the embedding space. Using a pre-trained foundation model, specifically the Arc2Face model, we synthesize identity-consistent images from these interpolated embeddings. We further enhance the realism of the generated images by modifying the intermediate CLIP space. The resulting dataset consists of various variations suitable for training a face recognition model.
The Face Recognition performance with Digi2Real dataset significantly improves over the DigiFace and achieves better performance than many other synthetic datasets.
Digi2Real dataset demonstrates superior performance over many alternatives, with particularly notable improvements on the IJB-B and IJB-C benchmarks, where it ranks sec- ond only to the DCFace dataset among the synthetic datasets. While performance on high-quality datasets like LFW remains comparable to other methods, our approach significantly outperforms many syn- thetic datasets on IJB-B and IJB-C, achieving verification rates of 70.14% and 75.80%, respectively.
New 🚀: The dataset is now available at the following link: https://www.idiap.ch/en/scientific-research/data/digi2real .
@article{george2024digi2real,
title={Digi2Real: Bridging the Realism Gap in Synthetic Data Face Recognition via Foundation Models},
author={Anjith George and Sebastien Marcel},
year={2024},
eprint={2411.02188},
url={https://arxiv.org/abs/2411.02188}, }