Retraction Watch logged over 10,000 cumulative retractions by early 2025. That number gets cited as proof the system works. It is actually proof the system is overwhelmed. The average time between publication and retraction runs roughly 3 years, and most retracted papers accumulate citations during that window. Downstream research builds on them. Clinical guidelines sometimes cite them. The correction arrives after the damage.

The self-correcting argument has a real point: exposure does happen. Press Gazette caught AI-fabricated content across 6 journalism outlets within 5 months in 2025, and retractions followed within weeks. That is genuinely fast. But journalism operates on a shorter cycle than peer-reviewed science, and the incentive structure is different. A retracted news story embarrasses an editor. A retracted paper rarely ends a career, and in some fields it barely slows one.

When the Detector Fails the Detected

The deeper problem is that AI-assisted fabrication has changed the detection math. Fabricated data used to leave statistical fingerprints: duplicated images, impossible standard deviations, suspiciously round p-values. Journals like PLOS ONE and Science have image-integrity software that catches some of this. But generative AI now produces synthetic datasets with realistic noise distributions. It writes methods sections that read like genuine lab work. The same dynamic that let AI-generated journalism evade detection at Wired in May 2025 applies to peer review, where editors are under more time pressure and have less institutional support for verification than most people assume.

A 2024 analysis of papers submitted to biomedical journals found that roughly 2% showed signs of AI-generated fabrication in figures or data tables. That sounds small. Biomedical journals publish around 1.5 million papers per year. Do the arithmetic.

The honest tension in my own argument: I cannot prove that fraud rates are rising faster than detection rates. The retraction count could reflect better detection, not more fraud. That possibility deserves weight. But the structural conditions, faster publication cycles, more journals competing for submissions, AI tools that lower the cost of fabrication, all point toward a system under increasing stress, not one finding equilibrium.

What the Newsroom Fixes Actually Teach Us

The journalism response in late 2025 included real-time video onboarding, doubled identity checks, and cross-outlet tip-sharing. Those are friction-adding measures. They slow the pipeline and make fabrication harder. Science has analogous tools that it systematically underuses: pre-registration of study designs, mandatory raw data deposit, registered reports where journals commit to publish based on methodology before results are known. Registered reports alone remove the publication-bias incentive that drives a significant share of p-hacking and selective reporting.

The journals that have adopted registered reports, around 300 as of 2024, represent a small fraction of active peer-reviewed publications. The NIH requires data management plans but does not uniformly enforce raw data sharing. Funders talk about open science; most grant cycles still reward novel positive results over rigorous null findings.

The self-correction optimists point to community norms and replication efforts like the Reproducibility Project. Those matter. But norms without enforcement are suggestions. The retraction system catches fraud after it has already shaped the literature. Pre-registration and mandatory data deposit catch it before. The difference is not procedural; it is whether bad science gets to spend 3 years building a citation record before anyone notices.

Journals should require raw data deposit as a condition of submission, not publication. Funders should tie renewal grants to compliance. That is a specific, enforceable ask. The 10,000th retraction is not a milestone. It is a backlog.