Establishing an evaluation metric to quantify climate change image realism

4Citations
Citations of this article
29Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

With success on controlled tasks, deep generative models are being increasingly applied to humanitarian applications (Nie et al 2017 Int. Conf. on Medical Image Computing and Computer- Assisted Intervention (Berlin: Springer) pp 417-25, Yanardag et al 2017 Deep Empathy). In this paper, we focus on the evaluation of a conditional generative model that illustrates the consequences of climate change-induced flooding to encourage public interest and awareness on the issue. Because metrics for comparing the realism of different modes in a conditional generative model do not exist, we propose several automated and human-based methods for evaluation. To do this, we adapt several existing metrics and assess the automated metrics against gold standard human evaluation.Wefind that using Fréchet Inception Distance with embeddings from an intermediary Inception-v3 layer that precedes the auxiliary classifier produces results most correlated with human realism. While insufficient alone to establish a human-correlated automatic evaluation metric, we believe this work begins to bridge the gap between human and automated generative evaluation procedures, and to generate more realistic images of the future consequences of climate change.

Cite

CITATION STYLE

APA

Zhou, S., Luccioni, A., Cosne, G., Bernstein, M. S., & Bengio, Y. (2020). Establishing an evaluation metric to quantify climate change image realism. Machine Learning: Science and Technology, 1(2). https://doi.org/10.1088/2632-2153/ab7657

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free