[ExI] AI Model Collapse

Stuart LaForge avant at sollegro.com
Fri May 30 15:27:56 UTC 2025


Since we have been talking about AIs recursively self-improving on the 
lists, I thought this was a pertinent article that could affect the 
timeline to AGI. There is a phenomenon called AI model collapse which 
occurs when AIs are trained on their own output. This produces an echo 
chamber effect which reinforces hallucinations, biases, and 
misinformation resulting in a degradation of output quality. Since these 
days, much of the output of AI gets put on the Internet and then new AI 
models get trained on the Internet, their training data becomes 
contaminated with AI output and can lead to AI model collapse. This is 
the informatic equivalent of biological inbreeding where deleterious 
mutations get amplified and reinforced in a genetic lineage resulting in 
all sorts of pathologies.

https://www.nature.com/articles/s41586-024-07566-y

Abstract
Stable diffusion revolutionized image creation from descriptive text. 
GPT-2 (ref. 1), GPT-3(.5) (ref. 2) and GPT-4 (ref. 3) demonstrated high 
performance across a variety of language tasks. ChatGPT introduced such 
language models to the public. It is now clear that generative 
artificial intelligence (AI) such as large language models (LLMs) is 
here to stay and will substantially change the ecosystem of online text 
and images. Here we consider what may happen to GPT-{n} once LLMs 
contribute much of the text found online. We find that indiscriminate 
use of model-generated content in training causes irreversible defects 
in the resulting models, in which tails of the original content 
distribution disappear. We refer to this effect as ‘model collapse’ and 
show that it can occur in LLMs as well as in variational autoencoders 
(VAEs) and Gaussian mixture models (GMMs). We build theoretical 
intuition behind the phenomenon and portray its ubiquity among all 
learned generative models. We demonstrate that it must be taken 
seriously if we are to sustain the benefits of training from large-scale 
data scraped from the web. Indeed, the value of data collected about 
genuine human interactions with systems will be increasingly valuable in 
the presence of LLM-generated content in data crawled from the Internet.
---------------------------------

Stuart LaForge






More information about the extropy-chat mailing list