SelfEval: Leveraging the discriminative nature of generative models for evaluation

17 Nov 2023  ·  Sai Saketh Rambhatla, Ishan Misra ·

In this work, we show that text-to-image generative models can be 'inverted' to assess their own text-image understanding capabilities in a completely automated manner. Our method, called SelfEval, uses the generative model to compute the likelihood of real images given text prompts, making the generative model directly applicable to discriminative tasks. Using SelfEval, we repurpose standard datasets created for evaluating multimodal text-image discriminative models to evaluate generative models in a fine-grained manner: assessing their performance on attribute binding, color recognition, counting, shape recognition, spatial understanding. To the best of our knowledge SelfEval is the first automated metric to show a high degree of agreement for measuring text-faithfulness with the gold-standard human evaluations across multiple models and benchmarks. Moreover, SelfEval enables us to evaluate generative models on challenging tasks such as Winoground image-score where they demonstrate competitive performance to discriminative models. We also show severe drawbacks of standard automated metrics such as CLIP-score to measure text faithfulness on benchmarks such as DrawBench, and how SelfEval sidesteps these issues. We hope SelfEval enables easy and reliable automated evaluation for diffusion models.

PDF Abstract
No code implementations yet. Submit your code now

Datasets


Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Benchmark
Visual Reasoning Winoground LDM-T5 (SelfEval) Text Score 29.00 # 74
Image Score 13.50 # 78
Visual Reasoning Winoground PDM-T5 (SelfEval) Text Score 28.25 # 76
Image Score 12.00 # 84
Visual Reasoning Winoground PDM-CLIP (SelfEval) Text Score 17.00 # 106
Image Score 14.00 # 73
Visual Reasoning Winoground LDM-CLIP (SelfEval) Text Score 22.75 # 92
Image Score 7.25 # 100
Visual Reasoning Winoground OCLIP (ViT-H/14) Text Score 30.75 # 62
Image Score 12.75 # 82
Visual Reasoning Winoground CLIP (ViT-L/14) Text Score 30.25 # 67
Image Score 8.0 # 96

Methods