FairGAN: Fairness-Aware GAN Framework
- FairGAN is a framework for creating synthetic datasets that balance realistic data utility with enforced demographic parity.
- Its dual-discriminator design and conditional MedGAN-inspired generator jointly optimize data utility, classification utility, and fairness.
- Empirical tests on the UCI Adult dataset show a 75% reduction in risk difference with only a minimal drop in downstream classification accuracy.
FairGAN is a framework for learning generative models that produce synthetic datasets both statistically similar to the original data and free of discrimination under group fairness notions such as demographic parity. Unlike naive de-biasing or traditional generative adversarial networks (GANs), FairGAN jointly enforces data utility, data fairness, classification utility, and classifier fairness, ensuring that downstream classifiers trained on synthetic data do not reproduce historical disparities (Xu et al., 2018).
1. Formal Problem Setting and Fairness Criteria
Given a dataset , with denoting unprotected features, a binary protected attribute (e.g., gender or race), and a binary outcome/label, FairGAN aims to learn a generator yielding a synthetic dataset that satisfies four desiderata:
- Data Utility: .
- Data Fairness: Statistical parity in the synthetic labels, i.e., .
- Classification Utility: A classifier trained on yields high accuracy on real .
- Classification Fairness: Classifier achieves on real .
FairGAN explicitly removes disparate impact by ensuring does not encode ; in practice, this is measured by the balanced error rate (BER) of a predictor , with BER close to 0.5 indicating minimal leakage.
2. Architecture and Learning Objectives
FairGAN uses a conditional MedGAN-inspired generator and two adversarial discriminators:
- Generator (): Accepts noise and protected attribute . It first produces a latent embedding via , then decodes to mixed discrete/continuous synthetic features using a pre-trained decoder (from an autoencoder). The output is with .
- Discriminator (Utility Critic): Receives and distinguishes real samples from generated.
- Discriminator (Fairness Critic): Receives and predicts . By minimizing 's ability to recover , FairGAN forces to remove undesirable correlations.
The full training objective is a dual-minimax game:
where:
is a hyperparameter that trades off data fidelity and fairness. is the standard conditional GAN loss; regularizes for independence between and (Xu et al., 2018).
3. Training Procedure and Control of Fairness–Utility Trade-off
Training involves two phases:
- Pre-train an autoencoder on the real for efficient mixed-type reconstruction.
- Alternately update , , , and again using minibatch stochastic gradient descent with Adam, sequentially applying the and -scaled losses to optimize for data utility and fairness.
Adjustment of interpolates between near-perfect data utility (, equivalent to a standard conditional GAN) and strong fairness (), allowing the practitioner to tune the fairness–utility trade-off according to application needs.
4. Empirical Evaluation and Results
Experiments on the UCI Adult dataset (48,842 instances, 57-dimensional one-hot encoded features, protected attribute: sex, label: income 50K) compare FairGAN against:
- SYN1-GAN: standard conditional GAN
- SYN2-NFGAN-I: GAN on with random reassignment of
- SYN3-NFGAN-II: two-discriminator GAN for , neglecting data matching
- SYN4-FairGAN: full objective with
Key metrics and representative results:
| Metric | Real | SYN1-GAN | SYN2-NFGAN-I | SYN3-NFGAN-II | SYN4-FairGAN |
|---|---|---|---|---|---|
| Risk diff | 0.1989 | 0.1798±0.0026 | 0.0025±0.0007 | 0.0062±0.0037 | 0.0411±0.0295 |
| BER (predict from ) | 0.1538 | – | – | – | 0.3862±0.0036 |
| Data utility ( joint dist) | – | 0.0198±0.0002 | – | – | 0.0208±0.0005 |
| SVM-Lin SYN2REAL Accuracy | 84.69% | 83.63±1.08% | – | – | 82.17±0.93% |
| SVM-Lin SYN2REAL Risk Diff | 0.1784 | 0.1712±0.0062 | – | – | 0.0461±0.0424 |
FairGAN reduces the risk difference in classifier predictions by approximately 75% at a cost of roughly 2% absolute decrease in accuracy for downstream tasks (Xu et al., 2018).
5. Extensions: Transfer and Reprogramming
A VAE-based reprogramming of FairGAN facilitates adaptation to new tabular datasets and tasks without retraining the entire model. A variational autoencoder’s decoder, pre-trained on the source data, acts as a fixed front-end for a new task-specific encoder and adversarial heads. This modularity enables rapid, resource-light transfer while maintaining the original targets of utility, fairness, and classifier performance. Trade-offs associated with this approach include increased hyperparameter sensitivity and possible convergence challenges in aligning fairness and accuracy on new domains (Nobile et al., 2022).
6. Limitations, Failure Modes, and Theoretical Constraints
FairGAN, in its original formulation, enforces only demographic parity; further group fairness notions (e.g., equalized odds, calibration) are not guaranteed and would require modifications such as additional adversarial heads predicting (Xu et al., 2018). Extension to multiple or non-binary protected attributes is not addressed in the initial proposal. Like other GAN-based approaches, FairGAN is prone to mode collapse and training instability, necessitating careful monitoring. Empirically, values of provide effective fairness–utility balances, but extreme weighting can degrade performance on either axis. In reprogrammed FairGANs, attaining perfect independence (50% accuracy) may not be achievable in all transfer scenarios without prohibitive utility loss, and the choice of latent dimensionality mediates a trade-off between representational capacity and privacy/sensitivity leakage (Nobile et al., 2022).
7. Significance and Impact on Fair Synthetic Data Generation
FairGAN is the first GAN-based approach for generating discrimination-free tabular data that effectively separates utility (realism) and fairness objectives via a dual-discriminator setup and a single interpretable trade-off parameter. It provides a pre-processing solution for de-biasing data such that both the generated datasets and the downstream classifiers inherit reduced disparate impact and disparate treatment. This has established FairGAN as a canonical reference point for subsequent fairness-aware generative modeling in structured data, with extensions such as VAE reprogramming further broadening its applicability (Xu et al., 2018, Nobile et al., 2022).