EN ES FR

Model Collapse

What happens when AI trains on AI output - and what it means for every organization using the same models on the same data.

In 2024, Shumailov et al. published in Nature what researchers had suspected: models trained on AI-generated data lose their tails. The distribution narrows. Rare events - the outliers, the edge cases, the genuinely novel - vanish from what the model can produce.

They called it model collapse. It is a property of the training process, not the model. And it compounds with each generation.

If your competitors use the same models on the same data,
you will converge to the same output.
Differentiation dies at the tail.

Distribution Narrowing Across Training Generations
Gen 0
Human
Full distribution. Rare events present at the tails.
Gen 1
AI + Human
Tails begin to thin. Rare cases underrepresented.
Gen 2
AI on AI
Distribution narrows further. Outliers mostly gone.
Gen 3
AI on AI
Only modal responses remain. Diversity collapsed.
Gen 4
AI on AI
Terminal convergence. Model produces near-identical outputs regardless of input variation.
THE MECHANISM

How Collapse Happens

The Paper

Shumailov et al. (Nature, 2024) trained successive generations of language models, each on outputs from the previous generation mixed with decreasing proportions of original human data. The finding was unambiguous: tails disappear early and irreversibly. The model's estimate of the data distribution shifts toward the mode - the most common, most average, most expected output. Low-probability events are first underweighted, then unlearnable, then unproduceable.

Why It Compounds

Each generation of model produces outputs biased toward the center of the previous generation's distribution. When that output becomes training data for the next generation, the bias amplifies. There is no self-correcting mechanism. The process is unidirectional. You cannot recover tail diversity by training on more AI output. You can only recover it by reintroducing human-generated data - data that hasn't been filtered through a model that already lost the tails.

The Internet Problem

The web is filling with AI-generated content at scale. The training sets for next-generation models will contain a growing proportion of AI-generated text, images, and code - produced by models that already exhibit some degree of collapse. Filtering is imperfect and increasingly difficult as generation quality improves. This is not a hypothetical future state. It is the current state of the data pipeline.

The model doesn't know it has lost the tails.
It produces what it knows with full confidence.
What it can't produce, it can't miss.
APPLIED: CONTENT

Marketing Goes Generic

What Happens

AI content tools trained on corpus data produce marketing copy, blog posts, and brand narratives that cluster around the most common patterns in their training set. The output is competent, readable, and indistinguishable from competitors using the same tool. The modal voice replaces the specific voice. At Gen 1, the difference is subtle. At Gen 3, every brand in the category sounds like a single averaged entity.

The Competitive Dynamics

  • Competitors using the same model on similar brief inputs converge toward similar outputs.
  • Editing AI output toward a brand voice is possible but requires the human judgment that AI was supposed to replace - eliminating the efficiency gain.
  • The most distinctive, highest-performing content tends to be the output that deviates most from the model's modal production. That content requires the human tail that the model has lost.
Organizational Consequence

The differentiation that made content marketing effective - a specific voice, a distinctive angle, a genuine perspective - is exactly what model collapse erodes. Brands that outsource content to AI without injecting authentic human distinctiveness will converge toward category average. Category average is not a competitive position. It is competitive neutrality - invisible to customers who have no reason to choose you over anyone else.

APPLIED: DECISIONS

Risk Misses Outliers

What Happens

AI decision-support tools - underwriting, credit scoring, investment screening, fraud detection - trained on historical data that includes AI-generated features or AI-labeled outcomes will exhibit collapse in their risk distributions. The outlier that doesn't look like past outliers is exactly the one that won't get flagged. The model has lost the capacity to recognize what it has never seen in adequate proportion.

The Tail Risk Problem

  • Financial tail risk - the events that happen rarely but catastrophically - lives at the tails of distributions. A collapsed model systematically underestimates tail risk.
  • Novel fraud patterns, new attack vectors, unprecedented market structures, and category-creating competitors all appear at the tails. A model that has lost its tails cannot recognize them.
  • The model's confidence does not decrease when it encounters tail events. It produces an average-case response with average-case confidence. The signal that something unusual is happening is absent.
Organizational Consequence

Organizations using AI decision-support for risk management are not just potentially exposed to unrecognized risks - they are structurally less capable of recognizing novel risk than the human analysts they replaced. Human analysts, for all their biases, are capable of noticing that something doesn't fit the pattern. A collapsed model is not. It fits everything to the pattern it has.

APPLIED: HIRING

Candidates Converge

What Happens

When candidates use AI to write applications and resumes, and when organizations use AI to screen them, both sides of the market are running through models trained on the same corpus of "effective" professional communication. The result is mutual optimization toward a single modal candidate profile. Applications that score well on AI screens are applications that look like other applications that scored well - which are applications that AI helped write.

The Talent Implications

  • Non-traditional candidates - career changers, autodidacts, people from industries that communicate differently - are penalized for not conforming to modal professional voice.
  • Organizations that rely on AI screening hire from a narrower effective talent pool than they believe. The screening is efficient. The selection is degraded.
  • If competitors use the same screening models, they hire the same modal profile. Workforce differentiation - the human capital advantage - collapses in parallel with the model.
Organizational Consequence

The genuinely unusual person - the one whose background doesn't fit the pattern but whose contribution would be transformative - is the person AI screening is least equipped to find. Organizations that want to hire for genuine distinctiveness cannot outsource that judgment to a model that has lost the capacity to recognize distinctiveness.

APPLIED: INNOVATION

Only Incremental Survives

What Happens

AI tools used in innovation processes - ideation, concept development, market research, competitive analysis - produce outputs biased toward the center of their training distribution. That center is the accumulated record of what has been done before. Genuinely novel ideas are, by definition, not well-represented in historical data. A model that has lost its tails cannot generate them, and cannot recognize them when a human proposes one.

The Innovation Dynamic

  • AI-assisted R&D, product ideation, and strategy development produces ideas that score well on feasibility metrics because they resemble ideas that were feasible before.
  • Radical ideas - the ones that break category assumptions, create new markets, or eliminate incumbents - score poorly on feasibility metrics because they don't look like anything that exists.
  • Organizations that filter innovation through AI scoring are not just less likely to discover radical ideas - they are actively selecting against them at scale.
  • Competitors using the same models converge on the same innovation roadmap. First-mover advantage disappears when everyone's AI is pointing at the same next move.
Organizational Consequence

An industry where all competitors use AI for innovation support will produce incremental innovation at industrial scale. The differentiated bet - the thing that looked unlikely but turned out to matter - is exactly what model collapse makes invisible. The organizations that win at category creation will be the ones that retained the human judgment to see what the model cannot.

You didn't lose the tail in a single moment.
It narrowed one generation at a time,
each time you fed the output back in.

Sources

Shumailov, I. et al. - "AI models collapse when trained on recursively generated data." Nature 631, 755–759 (2024) · Original observation: "The curse of recursion: Training on generated data makes models forget." arxiv:2305.17493 · Princeton AI Reliability Framework - arxiv:2602.16666 · Scar Taxonomy - unmake.it, on organizational knowledge lost to optimization