Synthetic Data Is Not About Replacing Reality. It Is About Questioning It.
Source: Dev.to
The Hidden Problem With Real‑World Data
We often talk about real‑world data as if it is neutral. It is not.
- Hiring data reflects decades of unequal access to education, employment, and opportunity.
- Healthcare data reflects who was diagnosed, who was believed, and who was ignored.
- Behavioural datasets reflect cultural norms and economic pressures.
When AI systems are trained purely on historical data, they do not learn fairness; they learn patterns—many of which are shaped by inequality. This is not a philosophical argument; it is a statistical one.
What Synthetic Data Actually Is
Synthetic data is artificially generated data that mimics the structure and statistical properties of real datasets without representing real individuals.
- It is not created for humans to read.
- It is created for systems to learn from, or to be tested against.
Examples
- Synthetic CVs are not meant to apply for jobs.
- Synthetic patient records are not meant to describe real people.
- Synthetic handwriting samples are not meant to replace human writing.
They exist to allow experimentation without harm.
Synthetic Data as a Controlled Lens
One of the most powerful properties of synthetic data is control. In the real world, you cannot ethically do the following:
- Take a job applicant.
- Change only their name, or their age, or a single line mentioning a disability.
- Then re‑run the application.
With synthetic data, you can.
Research on synthetic CV generation for fairness testing shows how artificial applicant profiles can be created where all variables are held constant except one. This allows researchers and practitioners to observe how automated hiring systems respond to specific demographic changes without involving real candidates or breaching privacy obligations (Saldivar, Gatzioura, Castillo, 2025).
When outcomes change under these controlled conditions, bias becomes visible—not as an accusation, but as observable behaviour.
Lessons From Healthcare and Rare‑Disease Research
Some of the most mature work on synthetic data comes from healthcare. In rare‑disease research, data is scarce, sensitive, and heavily regulated; sharing real patient records is often impossible.
- Privacy‑preserving synthetic data generation shows how generative models can create realistic patient profiles that allow analysis, model training, and collaboration without exposing personal information (Mendes, Barbar, Refaie, 2025).
These studies also highlight an important point: Synthetic data reflects the quality of the data it is generated from. If the original dataset is biased or incomplete, the synthetic data will inherit those weaknesses. This lesson transfers directly to hiring systems—synthetic data is not automatically fair; it must be designed with intent.
Why Representation Matters More Than Volume
Handwriting‑recognition research provides another insight. Some languages and writing styles are poorly represented in public datasets, causing models to perform well for some populations and poorly for others.
Large‑scale synthetic datasets are often required to capture enough variation for models to generalise properly, especially when real data is limited (Pham Thach Thanh Truc et al., 2025).
Takeaway: If certain groups are missing from the data, the system will struggle with them. This applies to CVs, medical records, and any system that interacts with human diversity.
What Robotics Teaches Us About Synthetic Worlds
Robotics offers a useful warning. In robotic learning, simulation is widely used because collecting real‑world data is expensive and slow. However, research on robotic bin‑packing shows that systems trained only in idealised synthetic environments often fail when deployed in real conditions (Wang et al., 2025).
Why? Because reality is messy:
- Objects behave unpredictably.
- Lighting changes.
- Constraints shift.
The same principle applies to synthetic data used for fairness testing. If synthetic CVs are too clean, too linear, or too idealised, fairness evaluations become misleading. Real careers are rarely neat—people change paths, take breaks, move countries, and care for others. Synthetic data must reflect this complexity to reveal meaningful bias.
Synthetic Data Does Not Eliminate Bias Automatically
Synthetic data does not fix bias on its own. Generative models learn patterns; they do not understand ethics or social context. If historical data encodes inequality, a naïve synthetic generator will reproduce it.
Recent research emphasises the need for constraints, validation, and domain knowledge when generating synthetic datasets, particularly in sensitive domains such as healthcare and employment (Mendes et al., 2025).
Synthetic data is a tool. Fairness depends on how it is used.
Why Synthetic Data Forces Honesty
Synthetic data removes excuses. When systems can be tested under controlled conditions, bias can no longer hide behind noise or complexity.
- If a hiring model behaves unfairly when only one variable is changed, the issue is structural.
- Synthetic data does not accuse; it reveals.
And that is precisely why it matters.
Looking Ahead
Synthetic data is often described as artificial, but its impact is real. It shapes how we:
-
Test AI systems.
-
Protect privacy.
-
Detect bias.
-
Imagine fairer alternatives.
-
Used carelessly, it can reinforce historical inequality.
-
Used thoughtfully, it can help us challenge it.
-
Synthetic data is not about replacing reality.
-
It is about questioning the systems we build from it.
References
- Saldivar, J., Gatzioura, A., & Castillo, C. (2025). Synthetic CVs to Build and Test Fairness‑Aware Hiring Tools. ACM Transactions on Intelligent Systems and Technology.
- Mendes, M., Barbar, F., & Refaie, A. (2025). Synthetic Data Generation: A Privacy‑Preserving Approach to Accelerate Rare Disease Research. Frontiers in Digital Health.
- Pham Thach Thanh Truc et al. (2025). HTR‑ConvText: Leveraging Convolution and Textual Information for Handwritten Text Recognition. arXiv preprint.
- Wang, Z. et al. (2025). RoboBPP: Benchmarking Robotic Online Bin Packing with Physics‑Based Simulation. arXiv preprint.
- MIT Technology Review – What synthetic data is and why it matters for AI
- Nature News and Comment – How artificial data could help address bias in AI
- OECD AI Policy Observatory – Fairness, transparency, and accountability in AI