The potential of synthetic image generation in machine learning applications has a myriad of both benefits and drawbacks. Synthetic datasets are valuable when real datasets are costly to acquire or difficult to share. Although synthetic imagery has improved in quality over the years, automated evaluation of the synthetic data remains challenging. For example, the Fréchet inception distance does not always correlate with perceptual quality and the perceptual path length operates at the model level rather than the data level. In this work, we propose a new evaluation metric that both correlates with perceptual quality and operates at the data level so that it can function on datasets from any domain. We do this by mapping high dimensional images into a lower dimensional, perceptually disentangled embedding space and computing the distance between distributions of embeddings within this space. We utilize a convolutional autoencoder trained with a linear combination of pixelwise and perceptual losses to perform the mapping and use existing metrics to measure the distance between the distributions of embeddings. We demonstrate efficacy of this metric on the CIFAR-10 dataset.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.