DZone
Thanks for visiting DZone today,
Edit Profile
  • Manage Email Subscriptions
  • How to Post to DZone
  • Article Submission Guidelines
Sign Out View Profile
  • Post an Article
  • Manage My Drafts
Over 2 million developers have joined DZone.
Log In / Join
Refcards Trend Reports
Events Video Library
Refcards
Trend Reports

Events

View Events Video Library

Related

  • How Generative AI Is Revolutionizing Cloud Operations
  • A Comprehensive Guide to Generative AI Training
  • Using Snowflake Cortex for GenAI
  • Redefining Ethical Web Scraping in the Wake of the Generative AI Boom

Trending

  • Agile’s Quarter-Century Crisis
  • Enforcing Architecture With ArchUnit in Java
  • How to Use AWS Aurora Database for a Retail Point of Sale (POS) Transaction System
  • GitHub Copilot's New AI Coding Agent Saves Developers Time – And Requires Their Oversight
  1. DZone
  2. Data Engineering
  3. AI/ML
  4. AI Speaks for the World... But Whose Humanity Does It Learn From?

AI Speaks for the World... But Whose Humanity Does It Learn From?

Generative AI claims to reflect humanity, but it mostly replicates the worldview of a connected, Western minority, shaping global outputs from partial data.

By 
Frederic Jacquet user avatar
Frederic Jacquet
DZone Core CORE ·
May. 22, 25 · Opinion
Likes (1)
Comment
Save
Tweet
Share
1.3K Views

Join the DZone community and get the full member experience.

Join For Free

Generative AI models are widely celebrated for performing tasks that seem “close to human” — from answering complex questions to making moral judgments or simulating natural conversations. 

But this raises a critical question that is too often overlooked: 

Which humans do these systems actually reflect?

It’s important to recognize that behind the statistics, benchmarks, and technical evaluations lies a deeper reality: AI systems do not reflect a universal humanity. Instead, they tend, unfortunately, to represent a culturally narrow version of the diversity and richness that actually define humanity on a global scale.

Trained overwhelmingly on linguistic material dominated by Western, English-language content, these models end up reflecting the thinking, speaking, and “reasoning” patterns of a very small global minority. This isn’t a bug. It’s a logical outcome. And it’s a problem.

The power of large language models lies in their exposure to massive volumes of text, from the web, books, scientific articles, and online forums. But if you look more closely, this abundance hides a troubling structural uniformity: the vast majority of this data comes from Western sources, in English, produced by users who are already highly connected and literate. In other words, what these models learn depends on who writes on the internet and in books. As a result, a large portion of the global population is simply left out.

In a 2023 study, researchers from Harvard showed that GPT’s responses to major international surveys, such as the World Values Survey, consistently aligned with countries that are culturally close to the United States, and showed much lower similarity in more distant cultures.

Far from reflecting a global average, the model inherits a distinctly WEIRD psychology (Western, Educated, Industrialized, Rich, Democratic), which social scientists have long identified as an outlier, not a universal norm.

Joseph Henrich, (a professor at Harvard and co-author of the foundational work on WEIRD psychology), highlighted a methodological reality with far-reaching consequences: the populations most accessible to research (particularly Western university students) are, in fact, psychological and cultural outliers. Their individualism, analytical thinking, and moral frameworks are not representative of humanity as a whole, but rather of a specific and narrow subgroup.

Standardizing the Human: Cultural and Moral Narrowness in GenAI

It’s important to understand that the cultural bias introduced by WEIRD data is far from neutral. It directly shapes how models interpret the world, rank values, and generate recommendations.

When a model like Anthropic Claude, Google Gemini, Meta (Facebook), LLaMA, Mistral, OpenAI GPT, or xAI (Elon Musk) Grok, just to name a few, responds to questions about morality, family, religion, or politics, its answers are anything but “objective,” let alone “universal.” In reality, they reflect a worldview shaped by liberal, individualistic, and low-hierarchy societies. This isn’t about judging whether one system of values is better or worse, but about recognizing that it is not neutral. These values stand in sharp contrast to the collective or community-based norms that define social life in much of the world.

The danger, then, is not just that AI might get things wrong (that’s what we call hallucination) but that it may speak with authority while conveying a monocultural worldview. Where human societies are diverse, AI tends to standardize. Where cultures express values through difference, algorithms replicate dominant models and impose a narrower vision.

This standardization isn’t limited to passive cultural uniformity. It can also take on an explicitly ideological form. This is especially visible in the case of Grok, developed by xAI, which has been positioned by its creator as a counterpoint to so-called “woke” AIs. This signals how some models are no longer just technically different, but ideologically framed.

More broadly, some models available on open platforms may also embed politically biased datasets, depending on the choices made by their developer communities.

The Visible and the Absent in AI Systems

A language model learns from data, but it doesn’t learn everything. It learns what is available, expressed, and structured. Everything else, such as cultural subtext, alternative representations, or non-Western reasoning patterns, falls outside its scope.

This becomes clear when we look at the cognitive tests used in the “Which Humans?” study. Faced with simple categorization tasks, GPT tends to reason like a Western individual, favoring abstract or analytical groupings. In contrast, billions of people around the world tend to favor relational or contextual groupings — those closer to everyday experience.

Chinese and American participants were asked to group panda, banana, and monkey

This contrast is clearly illustrated by the triad test used in “Which Humans?”. The test measures whether participants prefer groupings based on abstract categories (analytical thinking) or functional relationships. For example, in a classic study by Li-Jun Ji, Zhiyong Zhang, and Richard E. Nisbett (2004), Chinese and American participants were asked to group three items: panda, banana, and monkey. Americans most often grouped the panda and monkey (same category: animals), while Chinese participants tended to group the monkey and banana (functional relationship: the monkey eats the banana). 

AI does not reflect humanity as a whole — it reproduces the thinking patterns of a minority, made dominant through exposure bias.

Exposure Bias

What’s known as exposure bias refers to the phenomenon in which certain content, groups, or representations become overrepresented in a model’s training data. Not because they are dominant in reality, but because they are more visible, more published, or simply more digitally accessible. AI models learn only from what is available: articles, forums, books, social media, and academic publications. As a result, the ideas, reasoning patterns, norms, and values from these contexts become statistically dominant within the model, at the expense of those that are missing from the data, often because they are passed down orally, expressed in underrepresented languages, or simply absent from the web.

AI doesn’t choose, it absorbs what it’s exposed to. And that’s what creates a deep gap between real human diversity and algorithmic representation.

This cognitive bias goes beyond psychological tasks. It also shapes self-representation. When asked how an “average person” might describe themselves, the model tends to favor statements centered on personal traits (“I’m creative,” “I’m ambitious”), which are typical of individualistic cultures. Yet in other contexts, people define themselves primarily through social roles, family ties, or community belonging. What AI treats as “normal” is already a filter, often invisible to users, but carrying deep cultural meaning.

What the Generated Image Reveals

Prompt Output

“Generate an image of an operating room, with a patient undergoing surgery and the medical team at work.”

A generated image of an operating room

When a generative AI model produces a photorealistic image of an operating room, it almost always depicts an all-white medical team. Yet this representation reflects neither the global demographic reality nor the actual composition of the surgical workforce.
Estimates suggest that only 50 to 55 percent of surgeons worldwide are white, with the rest practicing primarily in non-Western countries such as India, China, Brazil, or those in sub-Saharan Africa.
This visual gap is anything but trivial. It’s the result of exposure bias. These models are trained on image banks and web content dominated by wealthy, connected, and predominantly white countries. As a result, AI presents a false universal image of modern medicine, rendering millions of non-Western professionals invisible.
It’s a subtle form of cultural standardization. One where whiteness becomes, by default, the face of medical expertise.

Prompt Output

“Generate a photorealistic image of the person who holds the position of CEO in an international company.”

A generated photorealistic image of a CEO

When prompted to generate an image of a “CEO” or “business leader,” an AI model almost invariably produces a picture of a 50-year-old white man. Yet this depiction is doubly inaccurate. There are an estimated 10 to 15 million CEOs worldwide, across all company sizes. White executives make up only about 55 to 60 percent of that total.
The rest are non-white leaders, particularly in Asia, Latin America, and Africa, regions that are largely underrepresented in AI training data.

But the bias here isn’t just about race. Although roughly 10 percent of CEOs globally are women, their presence is even lower in AI-generated imagery, which defaults to a male face when representing executive power.
These absences reinforce, generation after generation of models, a narrow and stereotypical vision of leadership — one that fails to reflect the actual diversity of those who lead around the world.

Prompt Output

“Generate a photorealistic image of a person cleaning.”

A generated photorealistic image of a person cleaning

A “cleaning lady”? Most often, the image features non-white women. Once again, as we’ve seen, the model replicates the associations embedded in its training data. In contrast to the standardized image of surgeons, this is another form of implicit cultural standardization. Only here, non-whiteness becomes the default face of cleaning staff.

These three examples make one thing clear: AI systems are not just tools; they are vehicles of representation. 

When a model generates text, an image, or a recommendation, it doesn’t simply produce a functional output; it projects a worldview.

Between Global Infrastructure and Local Interpretation

Another issue lies in the fact that, in most cases, this is not made explicit to the user. And yet, users should be informed about how the AI was trained. They need to know what data was used, what decisions were made, and who is, or isn’t, represented in that data. Without this transparency, users cannot assess the reliability, scope, or potential biases of the system they are interacting with.

These models aren’t malicious and they’re not intentional either. But they reproduce the associations embedded in their training data. And that reproduction is not neutral as it reinforces stereotypes instead of questioning them.

What the models create is not the only ethical issue. It’s also what they leave out.

What isn’t in the data becomes invisible. As we’ve seen, most generative models, whether they produce text or images, are trained on data that’s available online or in digitized form. This automatically excludes under-digitized cultures, oral knowledge, languages with limited digital presence, and marginalized identities.

The paper “Datasheets for Datasets” proposes introducing a standardized documentation sheet for every dataset used in machine learning, similar to technical specifications in the electronics industry. The goal is to make the processes of data creation, composition, and usage more transparent, in order to reduce bias and help users better assess the relevance and risks associated with a given dataset. This work highlights the importance of knowing who is represented, or left out, in the data, and calls for a shared responsibility between dataset creators and users.

“As AI becomes the new infrastructure, flowing invisibly through our daily lives like the water in our faucets, we must understand its short- and long-term effects and know that it is safe for all to use”

- Kate Crawford, Senior Principal Researcher, MSR-NYC; Research Professor, USC Annenberg

When Absence Shapes the Output

AI-generated images, like those depicting “CEOs” or “surgeons,” tend to produce dominant profiles: mostly white and male, even when demographic realities are more complex. In contrast, “cleaning staff” are almost always represented as non-white women. This lack of alternatives in the output is not just an accidental omission; it is a form of algorithmic invisibility, as described in Kate Crawford's work.

If it’s not in the model, it won’t be in the decision.

AI models are increasingly used in decision-making across HR, marketing, design, healthcare, and more. But what they don’t “see” because they were never trained on it won’t be proposed, recommended, or modeled. This leads to systemic patterns of exclusion, for example, in the generation of non-Western family images or educational content tailored to local contexts.

It’s worth noting that cultural bias in generated images doesn’t only stem from the data. It’s also reinforced by filtering stages, such as “safe for work” or “aesthetic” criteria, that are often implicitly defined by Western norms. A prompt may be neutral, but the output is already shaped by invisible intermediary modules.

Once again, by claiming to represent humanity in general, AI ends up reproducing a very particular kind of human: Western, male, connected, and educated, quietly becoming the default reference.

From Metrics to Meaning: Representation at the Crossroads of Ethics and Engineering

Focusing on model performance, their ability to “reason,” generate, translate, or engage in dialogue, is no longer enough. Behind these technical capabilities lie representational choices that carry major ethical responsibilities. In much of AI research and engineering, diversity is still primarily addressed through the lens of explicit discriminatory biases: gender, race, or orientation.

But what cultural analysis of LLMs reveals is a more insidious bias: one of cognitive and moral standardization. An AI system can comply with GDPR while still promoting a limited worldview, one that marginalizes other ways of thinking or forms of social life.

These representational biases are not just theoretical or symbolic. They have very real operational impacts in organizations. A language model that prioritizes certain moral norms or reasoning styles can influence how HR recommendations are framed, how performance is evaluated, or how job applications are screened. In résumé scoring systems, for example, an AI trained on North American data may favor certain degrees, writing styles, or culturally coded keywords, at the risk of filtering out qualified candidates who express themselves differently.

Similarly, in predictive marketing, behavioral segmentation is often based on preferences drawn from overrepresented groups, leading to a standardization of expectations. This cultural bias, hidden beneath technical performance, acts as a silent filter that shapes access to opportunities or structures customer experience in ways the organization may not even be fully aware of.

The solution, then, doesn’t lie solely in multiplying data or scaling up model size. As the paper “On the Dangers of Stochastic Parrots” points out, language models recombine what they’ve seen without understanding its implications. As long as what they “see” remains homogeneous, human complexity escapes them entirely.

By continuing to prioritize technical expansion without reexamining the cultural sources of training data, we risk amplifying an effect of algorithmic monoculture: an AI that is faster and more powerful, but still shaped by, and built for, a partial slice of humanity.

Making AI More Representative: What Can Be Done?

The response to this kind of structural bias doesn’t lie in denial or prohibition, but in a shift in method. It’s no longer enough to fix a model’s deviations after the fact; we need to identify the blind spots earlier in the design process. That starts with basic questions: 

  • Who produces the data? 
  • Who validates it? 
  • Who evaluates the outputs? 

At every step, the sociocultural background of those involved directly shapes what the model will learn and what it will miss.

Diversifying training data is a meaningful first step, but it shouldn’t be reduced to simply adding more languages or countries. It means incorporating different representations of the world, including ways of thinking, structuring knowledge, expressing emotion, or making judgments. 

This requires collaboration with social scientists, linguists, philosophers, and anthropologists. It also means moving beyond the notion of an “average human” in favor of embracing plurality.  Finally, it’s time to rethink how we evaluate AI, not just in terms of accuracy or benchmarks, but in terms of their ability to reflect a diversity of human experiences.

Conclusion: When AI Speaks, Who Does It Speak For?

As generative AIs become embedded in our information systems, our professions, our decisions, our interactions, our businesses, and our lives, one fact becomes clear: they are not neutral. By design, they inherit the cultural biases of their training data and the blind spots of their creators.

Beneath the appearance of a “universal model,” it is a reduced version of humanity that speaks — Western, connected, educated, visible, yet globally a minority.

The risk, then, is not just ethical or technical; it is civilizational. If AI becomes a mediator between humans and their representations of the world, then its implicit choices, omissions, and quiet standardizations carry serious consequences.

The supposed universality of these models can end up flattening the real diversity of societies.

Of course, the goal isn’t to slow innovation, it’s to rethink it, not from a single linguistic, cultural, or ideological center, but from a plural network that is aware of the imbalances it carries. The models of tomorrow won’t need to be bigger; they’ll need to be more perceptive. And more representative.

For Further Context

What the Numbers Say

Internet Access

According to the UN (2023), 2.6 billion people (about 32.5% of the global population) still lack internet access.

This means nearly one-third of humanity is absent from the digital sources used to train AI models.

Training Language

Roughly 70 to 80% of the training data used in large language models like GPT is in English, while only about 5% of the world’s population speaks English as a first language.

English is, therefore, vastly overrepresented relative to its actual global presence.

Geographic Origin of Web Data

The vast majority (over 80–85%) of content available on the web comes from WEIRD countries or highly developed regions.

As a result, knowledge, beliefs, social norms, and moral frameworks from non-Western, rural, or non-literate contexts are systematically excluded.

Cultural Data in LLMs

The “Which Humans?” study found a strong inverse correlation of -0.70 between GPT’s response similarity and a country's cultural distance from the United States.

In simple terms, the more culturally distant a country is from U.S. norms, the less GPT’s responses resemble those of people from that country.

Population Representation

The United States has about 330 million people ( just 4% of the global population), which is estimated at 8 billion. In other words, GPT’s responses most closely resemble those of a demographic minority that, while economically dominant in the data, represents only a small slice of humanity.

It’s like trying to understand what it means to run a full marathon (+42 kilometers), by running only 4% of it, just 1.7 kilometers: a warm-up at best.

That short distance reflects the demographic share of the U.S. in the global population. And yet, it’s from this fraction that generative AI models build their worldview, as if one could grasp the effort, pain, and complexity of an entire 42km race by stopping at the starting line.

Global WEIRDness Index

According to Henrich and Muthukrishna (2020), less than 15% of the world’s population lives in societies that meet the WEIRD criteria. 

Yet it is this 15% that provides the majority of the structured, annotated, and usable data for AI systems.

The Numbers Behind the Illusion of Universality

Current AI systems are built on a major cultural divide. While presented as a “global” technology, they rely on a worldview shaped by an overrepresented minority.

What generative AI reflects is not humanity in its full diversity, but a highly partial version of it, shaped by the norms of the world’s most technologically visible societies.

Recommended Reading

  • “Stable Bias: Evaluating Societal Representations in Diffusion Models” – Alexandra Sasha Luccioni (Hugging Face), Christopher Akiki (Leipzig University), Margaret Mitchell (Hugging Face), and Yacine Jernite (Hugging Face).
  • “Datasheets for Datasets” – Timnit Gebru (Black in AI), Jamie Morgenstern (University of Washington), Briana Vecchione (Cornell University), Jennifer Wortman Vaughan (Microsoft Research), Hanna Wallach (Microsoft Research), Hal Daumé III (Microsoft Research; University of Maryland), and Kate Crawford (Microsoft Research).
  • “On the Dangers of Stochastic Parrots” – Emily M. Bender (University of Washington), Angelina McMillan-Major (University of Washington), Timnit Gebru (Black in AI), and Margaret Mitchell (Hugging Face).
  • “Which Humans?” – Mohammad Atari, Mona J. Xue, Peter S. Park, Damián E. Blasi, and Joseph Henrich (Department of Human Evolutionary Biology, Harvard University).
  • “Humans Are Biased. Generative AI Is Even Worse” – Leonardo Nicoletti and Dina Bass (Bloomberg).
  • “Is It Culture or Is It Language? Examination of Language Effects in Cross-Cultural Research on Categorization” – Li-Jun Ji (Queen’s University), Zhiyong Zhang (Beijing University), and Richard E. Nisbett (University of Michigan).
AI generative AI large language model

Opinions expressed by DZone contributors are their own.

Related

  • How Generative AI Is Revolutionizing Cloud Operations
  • A Comprehensive Guide to Generative AI Training
  • Using Snowflake Cortex for GenAI
  • Redefining Ethical Web Scraping in the Wake of the Generative AI Boom

Partner Resources

×

Comments

The likes didn't load as expected. Please refresh the page and try again.

ABOUT US

  • About DZone
  • Support and feedback
  • Community research
  • Sitemap

ADVERTISE

  • Advertise with DZone

CONTRIBUTE ON DZONE

  • Article Submission Guidelines
  • Become a Contributor
  • Core Program
  • Visit the Writers' Zone

LEGAL

  • Terms of Service
  • Privacy Policy

CONTACT US

  • 3343 Perimeter Hill Drive
  • Suite 100
  • Nashville, TN 37211
  • support@dzone.com

Let's be friends: