ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.22450
23
0

Position: All Current Generative Fidelity and Diversity Metrics are Flawed

28 May 2025
Ossi Räisä
B. V. Breugel
M. Schaar
    EGVM
ArXiv (abs)PDFHTML
Main:10 Pages
22 Figures
Bibliography:3 Pages
4 Tables
Appendix:22 Pages
Abstract

Any method's development and practical application is limited by our ability to measure its reliability. The popularity of generative modeling emphasizes the importance of good synthetic data metrics. Unfortunately, previous works have found many failure cases in current metrics, for example lack of outlier robustness and unclear lower and upper bounds. We propose a list of desiderata for synthetic data metrics, and a suite of sanity checks: carefully chosen simple experiments that aim to detect specific and known generative modeling failure modes. Based on these desiderata and the results of our checks, we arrive at our position: all current generative fidelity and diversity metrics are flawed. This significantly hinders practical use of synthetic data. Our aim is to convince the research community to spend more effort in developing metrics, instead of models. Additionally, through analyzing how current metrics fail, we provide practitioners with guidelines on how these metrics should (not) be used.

View on arXiv
@article{räisä2025_2505.22450,
  title={ Position: All Current Generative Fidelity and Diversity Metrics are Flawed },
  author={ Ossi Räisä and Boris van Breugel and Mihaela van der Schaar },
  journal={arXiv preprint arXiv:2505.22450},
  year={ 2025 }
}
Comments on this paper