Skip to main content

Benchmark Study

Why Raw Synthetic Data Fails Clinical AI

Commodity synthetic generators score 65-75% on Train-Synthetic-Test-Real benchmarks. We quantify the gap across clinical domains and demonstrate how hybrid correction closes it.

Published April 13, 2026Download PDF

Synthetic data is widely promoted as a solution to healthcare AI's chronic data access problem. Vendors claim that algorithmically generated patient records can replace real clinical data for model training, offering privacy protection and unlimited scale. This paper examines that claim against published evidence and finds it wanting.

Commodity synthetic data generators — including rule-based engines like Synthea and uncalibrated generative adversarial networks — consistently score between 65% and 75% on Train-on-Synthetic, Test-on-Real (TSTR) benchmarks. Models trained on their output underperform by 25–35% compared to models trained on real patient data. A comprehensive benchmarking study across MIMIC-III and MIMIC-IV found AUC drops of 0.063 to 0.268 depending on the generator and task.

We identify five systemic failure modes: distribution errors that misrepresent disease prevalence across cardiology, endocrinology, oncology, and pediatrics; correlation breakdown that flattens the clinical relationships between diagnoses, labs, and medications; temporal artifacts that impose artificial regularity on care timing; missing clinical context that strips away physician reasoning and narrative documentation; and privacy theater that offers the appearance of confidentiality without formal guarantees.

Stadler, Oprisanu, and Troncoso demonstrated in a landmark 2022 USENIX Security paper that synthetic data provides a "false sense of privacy" — their evaluation showed that synthetic data either does not prevent inference attacks or does not retain data utility, and often fails at both.

The 65–75% TSTR ceiling is the difference between a toy and a tool. Closing the gap to 94%+ requires addressing each failure mode systematically through statistical correction, clinical enrichment, temporal calibration, privacy engineering, and rigorous multi-layer validation. This paper presents published evidence for each failure mode and the five questions every hospital data team should ask synthetic data vendors.

Full paper available

Download the complete white paper with methodology details, references, and supplementary data.

Download Full Paper

Questions about our methodology?

We welcome collaboration with health systems, academic researchers, and AI teams.

Contact Us