The author cited research that demonstrates that model collapse can happen on a small scale.
The author also cited sources that a larger and larger portion of the web will be written by language models.
There are already studies showing that LLM generated text is less diverse than human generated text:
https://techxplore.com/news/2026-03-llms-creativity-ai-respo...
https://arxiv.org/html/2501.19361
The studies don’t show that the lack of creativity in LLMs is caused by model collapse or that the problem is getting worse.
But 1) we know they do this and 2) we know that training on synthetic data can cause model collapse.
The main phrase of the title "model collapse is happening" is untrue and not substantiated in the article - all the true statements in the article are about the hypothetical problem, warning of the bad consequences that would likely happen if makers of major models did something they aren't doing, but they aren't doing that because that is a known issue that they're avoiding. It's like writing an article "Foot shooting epidemic is happening" with a long, solid (and true!) proof that if you'll shoot yourself in the foot, it will indeed cause serious injury...
yes, so given the title one might expect cited research that model collapse IS happening, as per OP's point.
Except perhaps the link to article on the peer-reviewed paper that describes the problem in detail.
https://www.cs.ox.ac.uk/news/2356-full.html
> Researchers at Oxford and Cambridge published work on this back in 2023, showing how iterative training on synthetic data leads to progressive degradation.
It does not say that it is happening in production LLMs. It is a theoretical concern right now.
Also I think this article itself may be AI-generated.