Have a personal or library account? Click to login
Benchmarking Tabular Data Synthesis: Evaluating Tools, Metrics, and Datasets on Prosumer Hardware for End-Users Cover

Benchmarking Tabular Data Synthesis: Evaluating Tools, Metrics, and Datasets on Prosumer Hardware for End-Users

Open Access
|Dec 2025

Full Article

1 Introduction

Data is essential in many domains; however, it is often unavailable or restricted due to privacy concerns (European Parliament, 2023). Synthetic data is a possible solution, providing datasets that mimic real data while improving scalability, privacy, and new scenario simulations (Xu et al., 2019). This study focuses on tabular data synthesis (TDS), where models replicate real dataset characteristics such as column count, data types, distributions, correlations, and integrity constraints. Unlike other tabular generation methods that rely on schema or rule-based statistics (e.g., Bruno and Chaudhuri, 2005; Gray et al., 1994; Neufeld, Moerkotte, and Lockemann, 1993), TDS directly models real data properties.

Previous work introduced a taxonomy for TDS models (Figure 1; Davila et al., 2025) identified key use cases and challenges, such as handling class imbalance, missing values, and generating realistic samples without replicating input data when (i) augmenting datasets, (ii) ensuring privacy, or (iii) creating scenario-specific data. However, the assessment was limited to the reported results of each tool.

dsj-24-2037-g1.png
Figure 1

Taxonomy of TDS models (illustration taken from Davila et al. (2025, Fig. 1)).

Several benchmarks have been proposed to systematically evaluate TDS models across diverse domains. SynthRO (Santangelo and Others, 2025) is a dashboard-based framework used for health-related synthetic tabular data, assessing resemblance, utility, and privacy metrics with a focus on electronic health records. A prominent benchmark, Synthcity (Qian, Davis, and van der Schaar, 2023), offers a comprehensive Python library that evaluates fidelity, utility, fairness, and privacy across various data modalities and use cases. The SDGym benchmark, part of SDV (Patki, Wedge, and Veeramachaneni, 2016), compares some of the main TDS models based on correlation, data proximity, and computational efficiency.

Our benchmark builds on these previous works to create a comprehensive evaluation across six critical dimensions in TDS: handling dataset imbalance, data augmentation, managing missing values, ensuring privacy, assessing ML utility, and computational performance. The 15 datasets used cover the use cases of these dimensions. We present a comprehensive experiment with 13 state-of-the-art TDS tools, evaluating a wide range of models, including Sampling, Bayesian Networks (BN), Generative Adversarial Networks (GAN), Variational Autoencoders (VAE), Diffusion, Graph Neural Networks (GNN), Transformers, and Hybrid models, as shown in Figure 1.

Importantly, we focus on end-user scenarios, evaluating tools on prosumer hardware, where high-performance computing may be unavailable. This ensures practical insights. Our contributions are: (i) An evaluation framework for benchmarking TDS tools in diverse use cases. (ii) An in-depth comparison of leading TDS models, including Sampling, BN, GAN, VAE, Diffusion, GNN, Transformers, and Hybrid models. (iii) Practical insights for selecting the most suitable TDS tool based on specific requirements.

The paper is structured as follows: Section 2 defines the scope of the benchmark. Section 3 describes the experimental setup. Section 4 presents the results across the six dimensions. Finally, Section 4.9 shows the aggregated results of the benchmark, and Section 5 concludes and discusses future work.

2 Scope Definition

2.1 Scope of tools

The tool selection process began with a query of scientific databases, including IEEE Xplore, ScienceDirect, Google Scholar, SpringerLink, NeurIPS proceedings, and MDPI journals. We included tools designed for TDS, excluding those for image or text generation. Selection criteria required peer-reviewed publications in English, with full-text and code availability, and alignment with TDS models included in Figure 1. In this taxonomy, hybrid models are those that combine different TDS models, for example, a VAE with diffusion. Imputation and discriminative models were excluded due to the absence of state-of-the-art tools.

To ensure a manageable comparison, we included only vanilla versions or significant advancements of tools representing each model type, excluding minor variations. The initial process identified 37 tools. The final verification in April 2024 excluded tools lacking publicly available code, highly customized tools like KAMINO (Ge et al., 2021) and STaSy (Kim, Lee, and Park, 2023), as they required specialized configurations incompatible with our standardized benchmark, and we focused on PyTorch-based tools, leaving TensorFlow-based implementations for possible future extensions. For all tools except SMOTE, we used the author-provided code. For SMOTE, we used Python packages imbalanced-learn (Lemaître, Nogueira, and Aridas, 2017) and smogn (Cantalupo, 2021).

Obsolete versions were replaced by advanced implementations, such as CTAB-GAN+ over TableGAN and GANBLR++ over GANBLR. Time-series tools were also excluded, as they are covered in custom benchmarks (Ang et al., 2023). The final tools are presented in Table 1.

Table 1

13 tools chosen for the benchmark.

MODELTDS TOOL
SamplingSMOTE (Cantalupo, 2021; Lemaître, Nogueira, and Aridas, 2017)
Bayesian NetworksPrivBayes (Zhang et al., 2017)
GANCTGAN (Xu et al., 2019), CTAB-GAN+ (Zhao et al., 2021), GANBLR++ (Zhang et al., 2022)
VAETVAE (Xu et al., 2019)
Diffusion (DPM)TabDDPM (Kotelnikov et al., 2023)
Graph NNGOGGLE (Liu et al., 2023)
TransformerGReaT (Borisov et al., 2023), REalTabFormer (Solatorio and Dupriez, 2023), TabuLa (Zhao, Birke, and Chen, 2025)
HybridAutoDiff (Suh et al., 2023), TabSyn (Zhang et al., 2024)

2.2 Scope of datasets

The initial 37 tools used 114 datasets in their evaluations. We first removed duplicates and modified versions of the same dataset, then prioritized datasets with diverse sizes, domains, column types, and characteristics like distributions and skewness. Publicly available datasets from the UCI repository (Dua and Graff, 2017) and Kaggle (LLC, 2010) were favored. Excessively large datasets were also removed to ensure compatibility with the hardware setup (Section 3). The final selection of 15 datasets is listed in Table 2, where mixed columns follow the definition in (Zhao et al., 2021) as columns containing both categorical and numerical values.

Table 2
DATASETCOLUMN NUMBERROW NUMBERCATEGORICAL COLUMNSCONTINUOUS COLUMNSMIXED COLUMNSML TASK
abalone (Nash et al., 1994)84177170Regression
adult (Becker and Kohavi, 1996)1448842932Classification
airline (Banerjee, 2016)1050000820Regression
california (Nugent, n.d.)520433140Regression
cardio (Janosi et al., 1989)1270000750Classification
churn2 (BlastChar, 2017)1210000561Classification
diabetes (Kahn, n.d.)9768270Classification
higgs-small (Whiteson, 2014)29627511280Classification
house (Torgo, 2014)16227840160Regression
insurance (Kumar, 2020)61338330Regression
king (harlfoxem, 2016)19216137102Regression
loan (Quinlan, 1987)125000660Classification
miniboone-small (Roe, 2005)51500001500Classification
payroll-small (City of Los Angeles, 2013)1250000480Regression
wilt (Johnson, 2013)64339150Classification

2.3 Scope of metrics

The initial compilation of evaluation metrics included 71 metrics sourced from the publications of the 37 tools, as well as from TabSynDex (Chundawat et al., 2024) and Goncalves et al. (Goncalves et al., 2020). Also, we identified key purposes for synthesizing data, along with functional and non-functional requirements users can impose on a TDS tool (Davila et al., 2025). Based on these, we designed specific tests to evaluate the tools, which include 23 of the 71 metrics from the tool publications, across six evaluation dimensions. These tests are shown in Table 3.

Table 3
EVALUATIONCHALLENGEEVALUATION FOCUSMETRICS
Dataset ImbalanceEnsuring that the tool is able to capture the real column distributions, even though there are imbalances in the classesClass distribution alignmentContinuous: Wasserstein Distance, KS Statistic, Correlation Differences. Categorical: Jensen–Shannon Divergence, KL Divergence, Percentage Class Count Difference
Data AugmentationGuaranteeing that the synthetic data generated remains realistic and meaningfulSimilarity and meaningful variability of new data pointsContinuous: Wasserstein Distance, KS Statistic, Correlation Differences, Quantile Comparison. Categorical: Jensen–Shannon Divergence, Percentage Number of Classes Difference
Missing ValuesMaking certain the tools are able to capture the key characteristics of the real dataset, even if it includes different levels of missing valuesSimilarity to original distributionsContinuous: Wasserstein Distance, Quantile Comparison. Categorical: Jensen-Shannon Divergence, Percentage Class Count Difference, Percentage Number of Classes Difference
PrivacyEnsuring whether the tools can generate truly synthetic data points rather than replicating the original data, which could potentially expose sensitive informationResemblance of synthetic records and real data and anonymity levelsDistance to Closest Record (DCR), Nearest Neighbor Distance Ratio (NNDR)
Machine Learning UtilityEnabling effective ML training with synthetic dataSynthetic datasets used to train ML models for classification and regression tasksAccuracy, Area Under the Receiver Operating Characteristic Curve (AUC), F1 Score (micro, macro, weighted), Explained Variance Score, Mean Absolute Error (MAE), and R2 Score.
PerformanceEnsuring synthetic data is generated within reasonable time frames while minimizing computational resource usage and maintaining scalabilityMeasure the computational resource usage and time required for data generationCPU Usage, GPU Usage, Memory Usage, Total Runtime

This study does not include evaluations for differential privacy, temporal dependencies, text columns, inter-table correlations, or integrity constraints. Including these aspects would significantly broaden the range of use cases, blurring the focus of this benchmark. Instead, these topics are reserved for future work.

3 Experimental Setup

This section outlines the experimental framework created in this study, consisting of three components: experimental setup, dataset preparation, and benchmarking architecture, to ensure consistent and reproducible results. We use the configuration parameters specified by the authors of each tool, for example, number of epochs for training or the choice of Large Language Model.

All experiments were conducted on a Linux laptop with 32GB RAM, an Intel Core i9-12900H (12th Gen, 14 cores, 20 threads, 2.5GHz base), and an external NVIDIA RTX 4090 GPU (24GB VRAM) in a Razer Core X eGPU enclosure with a 1000W PSU. This setup reflects prosumer hardware at the top of the consumer market, ensuring performance results are reproducible and not dependent on high-performance computing clusters.

All real datasets were pre-processed for consistency across tools. Categorical columns were numerically encoded using One-hot encoder, and missing values removed (a separate test covers missing value imputation). Datasets were shuffled and split to avoid temporal or positional bias. To ensure reproducibility, Table 4 summarizes how numerical data were normalized in each experiment. We replicate the original pre-processing procedures described in each paper.

Table 4

Normalization of datasets in the original experiments of different TDS tools.

TOOLNORMALIZATION STRATEGY
PrivBayesOnly discrete columns, no normalization
CTGAN / TVAEMode-specific normalization applied to all X
GANBLR++Ordinal encoding for all columns; numerical treated as discrete
TabDDPMNormalization of complete X (per code in data.py)
GOGGLENo normalization (raw tensors from get_dataloader)
GReaTNo normalization; textual encoding of all columns
REaLTabFormerNumerical columns normalized into fixed-length, digit-aligned string tokens
TabuLaNo normalization; continuous values directly as text tokens
AutoDiffAll numerical X normalized (Stasy: min–max; Tab: Gaussian quantile).
TabSynZ-score normalization of all numerical X

The exact configuration for each of the datasets can be found in Experiments. The synthetic datasets were also pre-processed before plotting the results, to ensure differences came from tool performance, not data representation. Min–Max scaling (mean of zero, standard deviation of one) was applied to continuous columns, and categorical encodings were aligned with the original datasets. The scaler was applied on Xtrain only and then applied to Xtrain, Xval, and Xtest, since fitting on all X can cause data leakage.

The benchmark architecture consists of the six steps shown in Figure 2. Each experiment is defined by a dictionary specifying the tool name, dataset, target column, problem type (classification or regression), and column types (categorical, continuous, mixed, text). The main benchmark.py script creates a Conda environment for each tool based on the configuration file, starts a shell script to monitor CPU, GPU, and memory usage, trains the model, generates five synthetic datasets with runtool.py, and evaluates them using the metrics in Table 3. Results are saved in performance and privacy folders for analysis.

dsj-24-2037-g2.png
Figure 2

Overview of the benchmark architecture, which automates configuration and analysis, ensuring reproducibility and consistency across tools and datasets.

Reproducibility is ensured by organizing each tool’s folder to include the original author code (modified only for path corrections or package updates), a runtool.py script for dataset generation, and a requirements.txt file listing necessary packages.

4 Results and Discussion

This section presents the evaluation results for each test in Table 3. The six evaluation dimensions include what we consider the key challenges TDS tools must overcome to produce satisfactory synthetic data: (i) capturing real dataset characteristics with class imbalance and missing values, (ii) generating realistic data points, (iii) avoiding direct replication of existing data, (iv) preserving data quality for downstream applications, and (v) ensuring generation time and resource usage are practical for end users. These challenges are based on two main aspects: accurate replication of column distributions and preservation of inter-column correlations. We first discuss these aspects.

4.1 Correlations

Relationships between columns are essential for structural integrity and accurate downstream tasks like predictive modeling. We used the Correlation Difference metric, which we calculate as in Correlations, where we first choose randomly column pairs in the dataset (always the same pairs for all experiments), identify from the configuration what type of columns they are, and calculate their correlation. For continuous columns, we use Pearson’s correlation; for binary categorical, we use the Point-Biserial; and for all other columns, we use Spearman and Kendall. All of them use the functions from the SciPy library. For all experiments which converged, we calculate the difference between the correlations of the original dataset’s column-pairs and the correlation of the synthetic datasets.

Figure 3 shows correlation preservation, with scores normalized from zero (perfect preservation) to one (worst performance). Empty cells are cases where no dataset was generated, mainly due to: (i) PrivBayes not supporting continuous target columns (making it unsuitable for regression use cases) (Zhang et al., 2017), or (ii) tools running out of resources, reflecting typical end-user hardware constraints.

dsj-24-2037-g3.png
Figure 3

Heatmap showing the correlation difference for various TDS tools and use cases, where zero indicates perfect preservation of inter-column correlations, and one represents the maximum difference. Empty cells denote cases where no dataset was generated due to intentional resource constraints.

Only AutoDiff, REaLTabFormer, SMOTE, and TVAE converged for all use cases, with AutoDiff and REaLTabFormer consistently preserving correlations. GAN-based tools (CTAB-GAN+, CTGAN, GANBLR++) struggled with datasets having many columns, often failing to converge or preserve correlations. GOGGLE (Graph Neural Networks) did not converge with any of the datasets using prosumer hardware, further discussed in subsequent evaluations. Notably, SMOTE and TVAE show low correlation errors, despite their relative simplicity.

To ensure fair comparisons, PrivBayes and GOGGLE were excluded from further evaluations. The remaining seven datasets, where all tools converged (abalone, diabetes, insurance, churn2, wilt, adult, cardio), were used for analysis. PrivBayes was separately assessed for classification tasks without continuous target columns to evaluate Bayesian Networks-based tools.

4.2 Column distributions

Preserving column-wise distributions is critical in TDS and closely related to the Privacy vs. Utility trade-off (Park et al., 2018), which describes the challenge of generating useful synthetic data for its original task while protecting sensitive information. Accurate column-wise distributions ensure that the synthetic dataset retains the statistical properties of individual columns, maintaining its utility.

We compared synthetic and real column distributions using visual inspection and quantitative metrics to assess how the tool captures the original statistical characteristics. Section 4.6 then analyzes privacy, the second part of the trade-off. Figure 4 shows one of the distribution plots used for visual inspection, showing how different tools replicate column-wise distributions across datasets. Visual inspection offers an intuitive initial check of whether synthetic data captures key characteristics. This assessment was performed for all datasets and tools, with plots available in Github Repository.

dsj-24-2037-g4.png
Figure 4

Distribution comparison for Wilt, with one row per TDS tool. For continuous columns, real data is in blue and synthetic in red. For categorical columns, bars show class counts with real data in blue and synthetic in red. The five synthetic distributions occasionally overlap completely. GOGGLE collapsed and therefore shows exploding densities.

Figure 4 presents plots for continuous and categorical columns. For continuous columns, real distributions are shown in blue and synthetic in red, showing whether tools capture column characteristics without assuming simple distributions like the Gaussian. For categorical columns, bar graphs display class counts, with real classes in blue and synthetic in red, making it easy to compare the number of classes and class proportions. For example, one can easily identify how GOGGLE collapses and reaches extremely high density values.

While visual inspection is a helpful starting point, it is inherently subjective and may overlook subtle discrepancies. To ensure consistency and precision, we used quantitative metrics such as the Wasserstein Distance for continuous data and Jensen–Shannon Divergence for categorical data. These metrics provide objective, numerical evaluations, capturing subtle differences in distributions that visual methods might miss, such as slight shifts in central tendency or variability. Quantitative metrics are essential for scalable and reproducible benchmarking, with results detailed in the following sections.

4.3 Dataset imbalance

Dataset imbalance is a key challenge in synthetic data generation (Davila et al., 2025), and it occurs when certain classes significantly outnumber others, potentially biasing predictive models toward dominant classes. To evaluate how well tools address this challenge, we used the metrics Wasserstein Distance (WD), Kolmogorov–Smirnov (KS) Statistic, Jensen–Shannon (JS) Divergence, and Kullback–Leibler (KL) Divergence, as implemented in the Python package (SciPy, 2024a, b, c, d).

For continuous columns, WD measures the cost of transforming one distribution into another, indicating how well tools maintain class proportions without distorting statistical properties. KS statistic captures the maximum difference between cumulative distribution functions (CDFs), reflecting alignment between real and synthetic class distributions.

For categorical data, JS Divergence measures the similarity between real and synthetic class probability distributions, with smaller values indicating better alignment. KL Divergence evaluates information loss when approximating distributions, penalizing deviations from target class balance. Percentage Class Count Difference provides an intuitive measure of relative class count differences between real and synthetic datasets.

Table 5 presents the evaluation results, with zero as the ideal value for all metrics. Top results are bolded and underlined. Each metric was calculated across all datasets for each tool, repeated five times for statistical significance, averaged, and normalized using Min–Max scaling for comparison.

Table 5

Dataset imbalance evaluation for all tools, averaged across datasets and normalized using Min–Max scaling, highlighting the top-performing tools (SMOTE, REalTabFormer, and TabSyn).

TOOLWASSERSTEIN DISTANCEKS STATISTICCORRELATION DIFFERENCEJS DIVERGENCEKL DIVERGENCEPERCENTAGE COUNT DIFFERENCE
AutoDiff0.2180.1640.1460.2110.1340.154
CTAB-GAN+0.3240.4160.2650.3440.3350.117
CTGAN0.2860.4820.3080.3820.4110.528
GANBLR++0.6920.7480.5420.7140.6590.489
GReaT0.5760.6740.2090.6530.1890.246
REalTabFormer0.1130.2940.0410.1850.1380.002
SMOTE0.0630.0620.1760.0150.0190.198
TabDDPM0.3440.3250.0530.2940.2760.861
TabSyn0.0400.1280.3240.0940.1090.176
TabuLaMiddle0.5130.6410.1830.5880.2010.233
TVAE0.3170.5430.3270.4230.3460.414

The top-performing tool was SMOTE, followed by CTAB-GAN+ and TabDDPM. As expected, SMOTE performed well because it directly oversamples minority classes by interpolating between similar points, making it ideal for dataset imbalance. However, this can risk creating synthetic samples that are too similar or reinforce linear patterns, as presented in (Brandt and Lanzén, n.d.) and evidenced in the further dimensions of this benchmark. In contrast, CTGAN and GANBLR++ produce relatively lower metrics, possibly because GAN models can have difficulty capturing minority classes due to mode collapse, where the generator fails to learn the minor classes’ full distribution.

4.4 Dataset augmentation

Dataset augmentation is also a primary use case for synthetic data (Davila et al., 2025), where users expand datasets to improve model robustness and generalization by generating new data points that preserve the original dataset’s statistical properties while enhancing diversity.

To evaluate how well a tool augments the data, we used again WD, KS Statistic, and correlation differences for continuous columns, but additionally Quantile Comparison, and JS Divergence with Percentage Number of Classes Difference for categorical columns. Quantile Comparison ensures synthetic data reproduces the original data’s spread across quantiles, crucial for maintaining variability and improving model robustness. Percentage Number of Classes Difference measures whether the synthetic data retains the original number of classes, with lower values indicating better representation.

Table 6 summarizes the results, where lower values represent better performance. Top results are bold and underlined. Metrics were calculated across all datasets for each tool, repeated five times for statistical significance, averaged, and normalized using Min–Max scaling for comparison.

Table 6

Augmentation evaluation results for all tools averaged across datasets and normalized using Min–Max scaling. The top-performing tools: TabSyn, SMOTE, REalTabFormer, and TabDDPM.

TOOLWASSERSTEIN DISTANCEKS STATISTICCORRELATION DIFFERENCEQUANTILE COMPARISONJS DIVERGENCEPERCENTAGE NUMBER CLASSES DIFFERENCE
AutoDiff0.2180.1640.1460.1970.2110.028
CTAB-GAN+0.3240.4160.2650.2870.3440.045
CTGAN0.2860.4160.2650.2260.3820.014
GANBLR++0.6920.7480.5420.6700.7140.096
GReaT0.5760.6740.2090.5430.6530.245
REalTabFormer0.1130.2940.0410.0920.1850.021
SMOTE0.0630.0620.1760.0710.0150.126
TabDDPM0.3440.3250.0530.2590.2940.006
TabSyn0.0400.1280.3240.0660.0940.123
TabuLaMiddle0.5130.6410.1830.4650.5880.332
TVAE0.3170.5430.3270.2870.4230.069

The top-performing tools were TabSyn, SMOTE, TabDDPM, and REalTabFormer. Unlike class imbalance handling, effective augmentation requires generating a diverse set of new samples that not only maintain statistical fidelity but also introduce meaningful variation. This is particularly captured by the Quantile Comparison metric, which evaluates how well tools reproduce the distributional spread across quantiles, a key aspect for improving model generalization. TabSyn showed the top performance on this metric, suggesting it effectively balances variability and structure. SMOTE and REalTabFormer followed closely, indicating that both sampling-based and Transformer-based methods can successfully generate data that enhances diversity while preserving core statistical properties. In the classification-only assessment, PrivBayes achieved the highest Percentage Number of Classes Difference. Simpler tools like CTGAN and TVAE perform well on basic metrics but struggle with complex distributions, supporting what is seen in Figure 4.

4.5 Missing values

Handling missing values is a key challenge in synthetic data generation. Missing values can distort analyses and degrade model performance by introducing gaps in the data. We evaluated tool performance by introducing 5%, 10%, and 20% missing values into the datasets and assessing their ability to maintain the integrity of continuous and categorical variables. For continuous columns, we used WD, KS statistic, correlation differences, and Quantile Comparison. For categorical columns, we used JS Divergence, Percentage Class Count Difference, and Percentage Number of Classes Difference, which ensures class diversity is not artificially altered when missing values affect certain classes.

Only 10 tools were evaluated, excluding GOGGLE and GANBLR++ due to poor initial performance, and TabDDPM due to errors with missing data in the configuration files. A visual inspection, shown in Figure 5, highlights the effect of missing values on the distributions of the diabetes dataset (Kahn, n.d.). Continuous plots display real distributions in blue, with red, green, and orange for synthetic datasets generated with 5%, 10%, and 20% missing values. Categorical bar plots show real classes in blue, with synthetic classes in red, green, and orange. The first column, SkinThickness, follows a multimodal distribution.

dsj-24-2037-g5.png
Figure 5

Distribution comparison for the Diabetes dataset, with 5%, 10% and 20% missing values. The real data distribution is plotted in blue, and synthetic data in red, green and orange. Continuous columns as density plots and categorical columns as bar plots.

Tools, such as AutoDiff, REaLTabFormer, SMOTE, TabSyn, and TabuLa, replicated distributions with minimal deviations, regardless of missing values. In contrast, CTGAN struggled, with deviations increasing as missing values rose, and it failed to preserve class proportions. CTAB-GAN+ captured simpler distributions but missed the multimodal nature of SkinThickness.

Metrics were calculated across all datasets for each tool at 5%, 10%, and 20% missing values, with evaluations repeated five times for statistical significance. Results were averaged and normalized using Min–Max scaling for fair comparison. Table 7a, 7b, and 7c show the detailed results. SMOTE demonstrated the highest robustness, maintaining stable performance across all metrics and missing value levels. AutoDiff, REaLTabFormer, TabSyn, and TabuLa also maintained acceptable performance at 20% missing values, while other tools struggled with higher levels of missing data.

Table 7

Missing-values evaluation results normalized using Min–Max scaling.

TOOLWASSERSTEIN DISTANCEKS STATISTICCORRELATION DIFFERENCEQUANTILE COMPARISONJS DIVERGENCEKL DIVERGENCEPERCENTAGE COUNT DIFFERENCEPERCENTAGE NUM CLASSES DIFFERENCE
(a) 5% Missing Values
AutoDiff0.2490.1780.6210.1940.3990.0510.0890.000
CTAB-GAN+0.4120.4120.2170.4080.4200.5450.0190.093
CTGAN0.8290.9300.6690.8550.8460.8790.7220.061
GReaT0.7440.6810.5780.7310.8430.0910.4980.138
REalTabFormer0.2510.4020.1210.2780.4600.5600.0040.147
SMOTE0.0610.0720.1820.0620.0480.0510.3760.000
TabSyn0.0740.1120.4180.0520.1230.0300.0451.000
TabuLaMiddle0.3620.4380.0910.4010.5310.0180.4520.055
TVAE0.8300.8790.3780.8350.8420.4220.4780.186
(b) 10% Missing Values
AutoDiff0.3680.1820.6420.3190.4150.0740.0690.000
CTAB-GAN+0.7980.6050.4710.7990.6350.6690.1470.092
CTGAN0.7250.9630.5230.7530.8720.8750.7340.034
GReaT0.7530.7110.5830.7350.8870.0770.4950.167
REalTabFormer0.2690.4360.1080.2630.4690.5310.0040.165
SMOTE0.0670.0650.1810.0290.0430.0470.3680.000
TabSyn0.0720.0920.0000.0380.0670.0750.0420.000
TabuLaMiddle0.3480.3980.0980.3820.3840.0150.4370.097
TVAE0.7350.9840.5370.7890.8040.3990.4590.261
(c) 20% Missing Values
AutoDiff0.4050.1840.5890.3850.3860.0580.0810.000
CTAB-GAN+0.3780.4420.2880.4190.4200.3560.0310.092
CTGAN0.9111.0000.6620.9760.9250.9040.7320.042
GReaT0.7690.6780.5630.7820.8630.0710.5010.129
REalTabFormer0.2670.4350.1280.3360.4920.4330.0050.138
SMOTE0.0490.0820.1620.0710.0510.0490.3780.000
TabSyn0.5020.9720.2310.5330.8350.0610.0210.033
TabuLaMiddle0.3340.4110.1160.4060.4660.5200.4510.057
TVAE0.6470.8930.3030.6580.7500.3110.4790.148

4.6 Privacy

Another key aspect of the Privacy vs. Utility trade-off is assessing how closely synthetic data resembles real data while ensuring privacy. This is crucial when real datasets contain sensitive or personally identifiable information. Synthetic data offers an alternative to traditional anonymization methods like masking, which often cause significant information loss. The challenge is to generate data that is useful for downstream tasks without exposing individual records.

We evaluated privacy using Distance to Closest Record (DCR) (Hernandez et al., 2022; Park et al., 2018) and Nearest Neighbor Distance Ratio (NNDR) (Zhao et al., 2024). DCR measures the Euclidean distance between each synthetic record and its nearest real counterpart. Lower DCR values enhance utility but increase privacy risk, while higher values suggest better privacy at the cost of utility. NNDR compares the distance between a synthetic record and its closest real record to its second-closest synthetic record, with higher values indicating better privacy protection through greater dispersion.

Table 8 summarizes the results, where higher values indicate stronger privacy protection. Top-performing tools are bolded and underlined. Each metric was calculated across all datasets and tools, repeated five times for statistical reliability, averaged, and normalized using Min-Max scaling for comparison.

Table 8

Privacy results for all tools averaged across datasets and normalized using Min-Max scaling, highlighting the top-performing tools (CTAB-GAN+, REalTabFormer, and TabDDPM). The best score is one, and the worst is zero.

TOOLDISTANCE TO CLOSEST RECORDNEAREST NEIGHBOR DISTANCE RATIO
AutoDiff0.1620.847
CTAB-GAN+0.1390.870
CTGAN0.2300.844
GReaT0.0460.751
REalTabFormer0.2830.781
SMOTE0.0320.342
TabSyn0.2610.750
TabDDPM0.3290.810
TabuLaMiddle0.0620.438
TVAE0.1010.837

TabDDPM, CTAB-GAN+, and REalTabFormer performed best in privacy. Despite its strong performance in previous evaluations, SMOTE underperformed in privacy, generating data points too similar to the original, demonstrating a poor Privacy vs. Utility trade-off. In contrast, CTGAN achieved high privacy scores but struggled with column distributions and correlations, also demonstrating the trade-off.

Tools like PrivBayes allow configurable parameters such as the privacy budget (𝜖), which controls the trade-off between utility and privacy under Differential Privacy (DP). DP adds controlled noise to ensure that including or excluding any single record does not significantly affect outputs, offering strong privacy guarantees. However, evaluating DP mechanisms is beyond the scope of this benchmark.

4.7 Machine learning utility

Machine learning (ML) Utility evaluates how well synthetic datasets replace real ones in training ML models for classification and regression tasks. To assess ML Utility, we used nine classifiers and four regressors, as shown in Table 9. Models were tested on real data to measure predictive accuracy and robustness. The metrics included Accuracy, AUC, and F1 Score (micro, macro, weighted) for classification, and EVS, MAE, and R2 for regression (Pedregosa et al., 2011). Accuracy measures overall correctness, AUC assesses class separation, and F1 Score balances precision and recall. EVS captures target variability, MAE measures average error, and R2 indicates the proportion of variance explained by the model.

Table 9

Classifiers and regressors used for the ML utility evaluation.

CLASSIFICATION METHODSREGRESSION METHODS
Decision TreesBayesian Ridge Regression
Gaussian Naive Bayes (NB)Lasso Regression
K-Nearest Neighbors (KNN)Linear Regression
Linear Support Vector Machine (SVM)Ridge Regression
Logistic Regression
Multilayer Perceptron (MLP)
Perceptron
Random Forest
Radial Support Vector Machine (SVM)

Table 10 presents the results as differences between model performance on real and synthetic data. Zero means equal performance, positive values favor real data, and negative values favor synthetic data. For MAE, lower values are better, so its inverse is shown for consistency. While accuracy and F1 Micro appear similar, accuracy measures overall correctness, while F1 Micro accounts for class imbalance. Similarly, EVS focuses on variability, while R2 also penalizes systematic prediction errors.

Table 10

ML utility results showing the difference between the model’s performance trained with real datasets and trained with synthetic datasets. For all metrics, if the difference is negative, the model performed better with synthetic data than with real data. If the difference is positive, the model performed worse.

TOOLACCURACYAUCF1 MICROF1 MACROF1 WEIGHTEDEVSINVERSE MAER2 SCORE
AutoDiff–0.0720.010–0.070–0.009–0.0800.0380.0090.062
CTAB-GAN+–0.0520.074–0.0460.0710.0530.1760.1450.172
CTGAN0.0890.0460.0810.0530.0930.1930.4720.192
GANBLR++0.1180.1520.1210.1560.1370.5480.3420.546
GReaT0.0430.0510.0400.0700.0330.3950.3980.665
REalTabFormer–0.024–0.029–0.020–0.028–0.0230.003–0.0260.003
SMOTE–0.011–0.014–0.013–0.017–0.0160.1430.0280.145
TabDDPM0.019–0.0310.017–0.0230.0100.0340.1820.035
TabSyn–0.046–0.041–0.043–0.040–0.0780.018–0.0350.017
TabuLaMiddle–0.063–0.072–0.059–0.052–0.0800.2040.0660.011
TVAE–0.024–0.015–0.021–0.006–0.0250.2100.1590.346

Table 10 shows that AutoDiff, TabuLa, REalTabFormer, and TabSyn preserve ML utility better. In general, tools perform better in classification than in regression, likely because the discrete nature of classification targets is easier to learn. SMOTE presented a strong classification performance; however, it is closely similar to the real data, which suggests limited record variability.

4.8 Computational performance

Computational performance measures the resource usage and execution speed during synthetic data generation, assessing computational efficiency and practical usability across tools. Advanced tools, such as deep learning-based models, often rely on GPUs, while simpler methods like PrivBayes and SMOTE prioritize lightweight execution without specialized hardware. This comparison highlights whether complex tools justify their resource demands with superior results in other dimensions (e.g., dataset balancing, or ML utility).

Table 11 shows the results for CPU, GPU, and memory usage, as well as runtime, measured by the mean, max, and standard deviation across five runs for all tools and datasets.

Table 11

Performance results for all tools averaged across datasets. Top-performing tools are SMOTE and PrivBayes; REalTabFormer shows the longest runtime and high resource usage.

TOOLMEAN CPU (%)MAX CPU (%)MEAN MEMORY (%)MAX MEMORY (%)MEAN GPU (%)MAX GPU (%)RUNTIME (S)
AutoDiff142415167373240
CTAB-GAN+1828783456863
CTGAN135913253236572
GANBLR++27881619001790
GOGGLE45562324001106
GReaT922131477863714
PrivBayes1732181900106
REalTabFormer1037213928777682
SMOTE2538171700248
TabDDPM113619202031666
TabSyn172519221357317
TabuLaMiddle2533232376901097
TVAE129017221315465

PrivBayes is the most resource-efficient tool (106s runtime, 17% CPU, no GPU), while SMOTE (248s, 25% CPU, no GPU) is the best for mixed data. REalTabFormer has the highest runtime (7682s) with high memory (21% mean, 39% max) and GPU usage (28% mean, 77% max), reflecting its complex architecture.

GOGGLE and GANBLR++ show high resource demands due to graphical modeling, with GOGGLE using 45% CPU and GANBLR++ peaking at 88% CPU.

GAN models like CTAB-GAN+ balance efficiency and runtime (863s, 18% CPU, 7% memory), while Transformer models (REalTabFormer, TabuLa) are resource-intensive, with TabuLa reaching 90% GPU usage. Diffusion-based TabDDPM offers balanced performance (11% CPU, 19% memory, 666s runtime). Hybrid models TabSyn (317s runtime, 17% CPU, 19% memory) and AutoDiff (14% CPU, 15% memory) provide efficient resource usage with minimal GPU reliance, maintaining strong performance across evaluations.

4.9 Aggregated benchmark results and discussion

We aggregated the results per model type across the six evaluation dimensions. Metrics were normalized, averaged per tool, and re-normalized to a 1–10 scale, where 10 means best performance, and 1 represents poor performance or failed use cases.

We assigned equal weight to all metrics within each of the six dimensions, aiming to provide general, comparable conclusions across models. We acknowledge that alternative weighting could be applied, depending on the priorities of specific use cases. Figure 6 compares the different TDS model types, using these normalized scores to provide a consolidated view of their overall strengths and limitations.

dsj-24-2037-g6.png
Figure 6

Spider web diagram summarizing model performance across the six benchmark dimensions. The scores are calculated by aggregating each dimension and are scaled 1–10.

Sampling models like SMOTE excel in handling imbalance, augmentation, and missing values, with top resource efficiency, making them ideal for fast, lightweight data synthesis. However, their low variability affects ML utility and privacy, as they often generate data too close to the original. While this limits their use in privacy-sensitive or high-variability tasks, their simplicity and low computational demands make them accessible for users with limited hardware, especially for balancing and augmentation needs in large datasets.

Graph Neural Networks (GNNs) like GOGGLE and GANBLR++ perform poorly with our prosumer setup, often failing to converge without high-performance clusters. Their reliance on complex graphical structures introduces computational overhead, making them unsuitable for prosumer hardware. They were still shown in Figure 6 for completeness, but the dimensions where they were excluded were left blank, since the results are not comparable. This is also the case for PrivBayes, which shows that Bayesian Networks are highly efficient, with strong privacy and performance scores, but their limited ability to model complex relationships affects their results for dataset augmentation and balancing, making them better suited for privacy-prioritized applications.

GAN-based models such as CTGAN and CTAB-GAN+ deliver moderate performance, excelling in augmentation and imbalance handling but struggling with missing values and privacy. Despite these challenges, GANs generate realistic synthetic data, provided the data is complete and pre-processed. Diffusion models achieve consistently high scores across all dimensions, balancing privacy, utility, and performance, but require significant tuning and resources for optimal use. Finally, Hybrid and Transformer models stand out with high scores in augmentation, missing values, and ML utility. Transformers (e.g., REaLTabFormer, TabuLa) deliver excellent results but demand substantial computational resources, reflecting their complexity. Hybrid models (e.g., AutoDiff, TabSyn) combine strengths from multiple models, offering balanced, efficient performance suitable for diverse use cases, making them versatile tools for most TDS applications.

4.10 Discussion on hyperparameter tuning

The experiments above were carried out, as mentioned in Section 3, using the parameters tuned by the original authors, without performing additional hyperparameter tuning. We chose this approach to avoid researcher bias and to benchmark the tools in their out-of-the-box configuration.

However, complex TDS models benefit greatly when their hyperparameters are tuned for each dataset. Simple TDS tools, such as SMOTE and PrivBayes, require no hyperparameter tuning because they rely on fixed algorithmic procedures rather than parameterized learning processes. In contrast, GANs are well known to be unstable during training, meaning that key hyperparameters, such as learning rates, optimizer choice, gradient penalty, and batch size, can cause training to collapse or oscillate. (Arjovsky, Chintala, and Bottou, 2017; Mescheder, Geiger, and Nowozin, 2018) For this reason, GAN-based tools typically incorporate stabilization techniques, such as Wasserstein loss or spectral normalization, which should not be, or only very carefully, altered.

Diffusion models, on the other hand, are much more stable to train because their training is explicit likelihood-based, as shown in Figure 1. Yet, their effectiveness is sensitive to the chosen hyperparameters (Kotelnikov et al., 2023), such as the number of diffusion steps, learning rate, or noise schedule. As a result, hyperparameter tuning can significantly improve the performance of diffusion-based models. The situation is similar for Transformer-based tools, which depend heavily on their hyperparameters (Casola, Lauriola, and Lavelli, 2022), such as number of layers, hidden dimension, number of attention heads, or learning rate schedules.

This study acknowledges that a further benchmark including hyperparameter tuning would increase fairness toward certain models, such as Diffusion or Transformers, which are capable of producing superior results when properly tuned. Nevertheless, Diffusion, Transformer-based, and Hybrid models already achieved better performance in our benchmark with respect to correlations and distributions, which are the basis of synthetic data quality. In the aggregated results, they obtained the highest scores across all dimensions, except computational performance.

We expect that including hyperparameter tuning in future experiments will further increase the quality of the generated synthetic data, but at the expense of higher computational cost. Exploring this trade-off is left for future work.

5 Conclusion

Through the evaluation across six dimensions (dataset imbalance, data augmentation, handling missing values, privacy, ML utility, and performance), we gained valuable insights into the capabilities and limitations of individual tools and models for TDS. Sampling-based tools like SMOTE demonstrated exceptional performance in dataset imbalance, augmentation, and resource efficiency, but struggled with privacy and variability. Hybrid and Transformer models stood out as the most consistent performers across all dimensions, achieving high scores in utility and privacy but requiring substantial computational resources. In contrast, GNNs and GANs combined with graphical models often failed to converge under our setup, highlighting their unsuitability for environments without access to high-performance clusters.

Our findings also emphasize the trade-offs inherent in different TDS models. Diffusion models showed promise with high scores across evaluations but were complex to configure, while Bayesian Networks offered strong privacy protection and efficiency but limited utility in dataset imbalance and augmentation tasks. VAEs, although moderate in performance, serve as the basis for high-performing Hybrid models. GANs displayed versatility in augmentation and dataset imbalance but had moderate to low privacy and utility results, indicating the need for careful consideration in sensitive use cases.

Beyond the quantitative findings, we observed additional non-functional aspects that influence tool usability. Tools like TabDDPM and TabSyn, despite their strong performance, were challenging to implement and required significant effort to configure properly. Others, such as GANBLR++ and GOGGLE, had complicated requirements and dependencies, making them time-consuming to deploy effectively. These considerations, while not easily quantifiable, are critical when selecting a TDS tool, as they directly impact the practicality and adoption of these models in real-world scenarios. Overall, this benchmark provides a detailed roadmap for researchers and practitioners to navigate the landscape of TDS tools, aligning their choices with specific needs and constraints.

This benchmark is limited by the fact that we do not perform any hyperparameter tuning for complex models, such as Diffusion or Transformers, which may impact their effectiveness. Our future work includes experiments with hyperparameter tuning (Davila, Turaev, and Wingerath, 2025).

Appendices

Appendix A: Library Versions in the Experiments

For the sake of reproducibility and software versioning, we add the package requirements for each of the TDS tools used in this experiment.

Table 12

TDS tools and required packages.

TDS TOOLPACKAGE REQUIREMENTS
SMOTEpandas==2.2.2, numpy==2.0.0, scikit-learn==1.5.2, imbalanced-learn==0.13.0
PrivBayesdiffprivlib==0.6.3, dill==0.3.7, dython==0.6.8, joblib==1.2.0, lifelines==0.27.8, matplotlib==3.7.2, numpy==1.26.0, pandas==1.3.4, pyjanitor==0.26.0, pandas_flavor==0.6.0, scikit_learn==1.3.0, scipy==1.11.3, seaborn==0.13.0, thomas_core==0.1.3, synthetic-data-generation, torch, gpustat
CTGANtqdm==4.66.5, torch==2.1.0, numpy==2.0.0, pandas==2.2.2, scikit-learn==1.5.2, ctgan, joblib==1.4.2, rdt==1.7.0
CTAB-GAN+numpy==1.21.0, torch==1.10.0+cu113, torchvision==0.11.1+cu113, torchaudio==0.10.0+cu113, pandas==1.2.1, scikit-learn==0.24.1, dython==0.6.4.post1, scipy, gpustat, tqdm, -f https://download.pytorch.org/whl/torch_stable.html
GANBLR++ganblr
TVAEtqdm==4.66.5, torch==2.1.0, numpy==2.0.0, pandas==2.2.2, scikit-learn==1.5.2, ctgan, joblib==1.4.2, rdt==1.7.0
TabDDPMcatboost==1.0.3, category-encoders==2.3.0, dython==0.5.1, icecream==2.1.2, libzero==0.0.8, numpy==1.21.4, optuna==2.10.1, pandas==1.3.4, pyarrow==6.0.0, rtdl==0.0.9, scikit-learn==1.0.2, scipy==1.7.2, skorch==0.11.0, tomli-w==0.4.0, tomli==1.2.2, tqdm==4.62.3
GOGGLEchardet==5.1.0, cvxpy==1.1, dgl==0.9.0, geomloss==0.2.5, matplotlib==3.7.0, numpy==1.23.0, packaging==21.3, pandas==1.4.3, pgmpy==0.1.21, scikit-learn==1.1.1, seaborn==0.12.2, synthcity==0.2.2, torch==1.12.0, torch-geometric==2.2.0, torch-sparse==0.6.16, torch_scatter==2.1.0
GReaTdatasets≥2.5.2, numpy≥1.23.1, pandas≥1.4.4, scikit_learn≥1.1.1, torch≥1.10.2, tqdm≥4.64.1, transformers≥4.22.1, accelerate≥0.20.1
REalTabFormertorch, bandit≥1.6.2,<2.0, black~=22.0, build~=0.9.0, import-linter[toml]==1.2.6, openpyxl~=3.0.10, pre-commit≥2.9.2,<3.0, pylint≥2.5.2,<3.0, pytest-cov~=3.0, pytest-mock≥1.7.1,<2.0, pytest-xdist[psutil]~=2.2.1, pytest~=6.2, trufflehog~=2.1, twine~=4.0.1, pandas, datasets, scikit-learn, transformers, realtabformer
TabuLadatasets≥2.5.2, numpy≥1.24.2, pandas≥1.4.4, scikit_learn≥1.1.1, torch≥1.10.2, tqdm≥4.64.1, transformers≥4.22.1
AutoDiffnumpy==2.0.0, pandas==2.2.2, scikit-learn==1.5.2, scipy==1.10.1, torch==2.1.0, gpustat==1.0.0, psutil==5.9.4, tqdm==4.65.0, ipywidgets==7.8.5, jupyter==1.0.0, matplotlib==3.7.1
TabSynnumpy==2.0.0, pandas==2.2.2, scikit-learn==1.5.2, scipy==1.10.1, torch==2.1.0, icecream==2.1.2, category_encoders==2.3.0, imbalanced-learn==0.14.0, transformers==4.25.0, datasets==2.8.0, openpyxl==3.1.2, xgboost==1.7.5

Competing Interests

The authors have no competing interests to declare.

Author Contributions

  • Maria Davila was the main contributor to this work. She carried out the experiments and designed the methodology, playing a central role in the conception and execution of the study.

  • Benjamin Wollmer contributed to the design of the evaluation process and participated in the critical review of the proposed method, ensuring its robustness and relevance.

  • Fabian Panse was responsible for the technical correctness of the benchmark and implementation of the data synthesis tools, contributing significantly to the reproducibility and reliability of the results.

  • Wolfram Wingerath served as the supervising professor. He ensured the scientific validity of the methodology, evaluation, and writing process and provided guidance throughout the development of the work.

Language: English
Submitted on: Jul 7, 2025
Accepted on: Nov 19, 2025
Published on: Dec 9, 2025
Published by: Ubiquity Press
In partnership with: Paradigm Publishing Services
Publication frequency: 1 issue per year

© 2025 Maria Fernanda Davila Restrepo, Benjamin Wollmer, Fabian Panse, Wolfram Wingerath, published by Ubiquity Press
This work is licensed under the Creative Commons Attribution 4.0 License.