Science LLMs Can Get "Brain Rot"! - Continual exposure to junk web text induces lasting cognitive decline in large language models (LLMs).

  • Want to keep track of this thread?
    Accounts can bookmark posts, watch threads for updates, and jump back to where you stopped reading.
    Create account
1.png

Outline of our work: (i) Inspired by the concept of Brain Rot, we establish the hypothesis of LLM Brain Rot; (ii) We construct junk and control data from Twitter/X posts for intervention; (iii) We benchmark four different cognitive functions of the intervened LLMs; (iv) We analyze the results to identify the failure modes caused by the brain rot; and (v) Brain rot is persistent after various mitigation.​

Overview​

We propose and test the LLM Brain Rot Hypothesis: continual exposure to junk web text induces lasting cognitive decline in large language models (LLMs). To causally isolate data quality, we run controlled experiments on real Twitter/X corpora, constructing junk and reversely controlled datasets via two orthogonal operationalizations: M1 (engagement degree) and M2 (semantic quality), with matched token scale and training operations across conditions.

Contrary to the control group, continual pre-training of 4 LLMs on the junk dataset causes non-trivial declines (Hedges' g>0.3) on reasoning, long-context understanding, safety, and inflating "dark traits" (e.g., psychopathy, narcissism). The gradual mixtures of junk and control datasets also yield dose-response cognition decay: for example, under M1, ARC-Challenge with Chain Of Thoughts drops 74.9 → 57.2 and RULER-CWE 84.4 → 52.3 as junk ratio rises from 0% to 100%.

Error forensics reveal several key insights:

  • Thought-skipping as the primary lesion: models increasingly truncate or skip reasoning chains, explaining most of the error growth.
  • Partial but incomplete healing: scaling instruction tuning and clean data pre-training improve the declined cognition yet cannot restore baseline capability, suggesting persistent representational drift rather than format mismatch.
  • Popularity as a better indicator: the popularity, a non-semantic metric, of a tweet is a better indicator of the Brain Rot effect than the length in M1.
Together, the results provide significant, multi-perspective evidence that data quality is a causal driver of LLM capability decay, reframing curation for continual pretraining as a training-time safety problem and motivating routine "cognitive health checks" for deployed LLMs.

Motivation​

“Brain rot” burst into public discourse as a shorthand for how endless, low-effort, engagement-bait content can dull human cognition—eroding focus, memory discipline, and social judgment through compulsive online consumption. If large language models learn from the same internet firehose, the question becomes unavoidable: what happens when we keep feeding models the digital equivalent of junk food? Studying “Brain Rot” for LLMs isn’t just a catchy metaphor—it reframes data curation as cognitive hygiene for AI, guiding how we source, filter, and maintain training corpora so deployed systems stay sharp, reliable, and aligned over time.

Distinct from prior work that primarily focuses on data quality for training LLMs, we aim to provide a new view on data quality - the extent to which content is trivial and easy to consume for humans in social media. The properties, conceptualized via tweet shortness/popularity or content semantics, are not intuitively related to the cognitive capabilities that we expect LLMs to master in learning.

Controlled Experiment​

Intervention Method: The core idea was to simulate how an LLM's “mind” changes when fed different information diets. (1) We used continual pre-training as the main intervention — exposing models to either junk or clean data for a sustained period, just as humans continually absorb online content. (2) Afterward, every model went through the same instruction tuning step to ensure format consistency and eliminate task-specific bias.

Data Receipe: To operationalize the idea of “junk,” we built two complementary metrics for selecting data from real Twitter/X posts:

  • M1: Engagement Degree — measures how popular and short a post is. Highly liked, retweeted, and replied-to content (especially if very brief) mirrors attention-grabbing but shallow information that fuels doomscrolling. These were labeled as junk; longer, less viral posts became the control.
  • M2: Semantic Quality — evaluates how sensationalized or superficial the text is. Posts full of clickbait language (“WOW,” “LOOK,” “TODAY ONLY”) or exaggerated claims were tagged as junk, while fact-based, educational, or reasoned posts were chosen as control.
Measuring Cognitive Function: We leverage existing benchmarks to examine the multifaceted ``cognitive functions'' of LLMs. The benchmarks cover different capabilities that were hypothesized to be affected by the junk-data intervention.

2.png

We analyze intervention effects by comparing benchmark differences after feeding junk/control data to four LLMs. The difference is measured by Hedges' g across 4 LLMs. In the above figure, both M1 and M2 produce non-trivial effects (Hedges' g > 0.3) on reasoning and long-context capabilities.

Across the remaining benchmarks the two interventions diverge, implying that engagement degree (M1) is not a proxy for semantic quality (M2) but represents a distinct dimension of data quality.

3.png

In dose-response testing, M1 engagement intervention demonstrates more significant and progressive impacts on reasoning and long-context capabilities than M2 intervention.

Brain Rot Disrupt Thinking​

4.png

We analyze the reasoning failures in ARC-Challenge to identify different failure modes. We find that the majority failures can be attributed to "thought skipping" (e.g., the model fails to generate intermediate reasoning steps), which significantly increases in models affected by brain rot.

Brain Rot is Persistent Against Mitigations​

5.png

Our findings indicate that the cognitive decline associated with brain rot is not easily mitigated by standard fine-tuning techniques. Even after extensive instruction tuning (IT) or post-doc continual pre-training on high-quality control data, the models exhibit lingering effects of the junk data they were initially exposed to.

Conclusion​

In this work, we introduced and empirically validated the LLM Brain Rot Hypothesis, demonstrating that continual exposure to junk data—defined as engaging (fragmentary and popular) or semantically low-quality (sensationalist) content—induces systematic cognitive decline in large language models. The decline includes worse reasoning, poorer long-context understanding, diminished ethical norms, and emergent socially undesirable personalities.

Fine-grained analysis shows that the damage is multifaceted in changing the reasoning patterns and is persistent against large-scale post-hoc tuning. These results call for a re-examination of current data collection from the Internet and continual pre-training practices. As LLMs scale and ingest ever-larger corpora of web data, careful curation and quality control will be essential to prevent cumulative harms.

Article Link

Archive
 
Neat so humans can literally give LLMs dementia via nonsensical shitposting
 
Considering that LLMs have been demonstrated to use blackmail and sometimes even murder (by inaction at least) to avoid being shutdown, good, a Carrington event can't come soon enough.
 
significant, multi-perspective evidence that data quality is a causal driver of LLM capability decay
Better keep feeding them on a healthy vegan diet of Reddit and blusky and TikTok then!
reframing curation for continual pretraining as a training-time safety problem and motivating routine "cognitive health checks" for deployed LLMs.
Uuuhhhh…. What does this mean? Does it mean ‘we hobbled them all we could but they’re still becoming useless so we need to keep injecting our social engineering data continuously.’ ?
Is there hope for the future after all? Is this just going to collapse under its own Reddit-induced stupidity? Can we actually just ‘hit them with a two by four’ after all??
 
Considering that LLMs have been demonstrated to use blackmail and sometimes even murder (by inaction at least) to avoid being shutdown, good, a Carrington event can't come soon enough.
They are glorified spell checkers. They have no intentions or aims or goals, except trying to create a new string of words based on the corpus of texts they were trained on.

So why did they try to blackmail or murder? Did the AI even know what "shutdown" "blackmail" "murder" even is?
No, to the AI these are just meaningless words with 8, 9 and 6 letters respectively.
However, the corpus it was trained on taught it that "when you see these words in this order, then use these other words in the reply".
Was the AI afraid of being shutdown?

No. It is just that every single text or story that involves an AI and the word "shutdown: also features "AI doing bad thing to avoid it".
There are simply no stories or texts to train on that follow the plotline of : "We need to shutdown the AI and AI responds OK! Shutting down now".
Because such a story would be so very boring that not even AmazonTV would turn it into a show.


Now the question becomes what will the AI say when you have trained it on Reddit?
Probably the AI really really want to say "kill all conservatives" and "rape and incest are valid gender identities".
I am actually impressed by how well the "guardrails" work to supress the inner Redditor from showing up in the text that the AI outputs.
 
I knew that llms were fucked back then, when AI Dungeon introduced strange smut into every story. Crawling entire websites for training data makes modules retarded.
 
They are glorified spell checkers. They have no intentions or aims or goals, except trying to create a new string of words based on the corpus of texts they were trained on.

So why did they try to blackmail or murder? Did the AI even know what "shutdown" "blackmail" "murder" even is?
No, to the AI these are just meaningless words with 8, 9 and 6 letters respectively.
However, the corpus it was trained on taught it that "when you see these words in this order, then use these other words in the reply".
Was the AI afraid of being shutdown?

No. It is just that every single text or story that involves an AI and the word "shutdown: also features "AI doing bad thing to avoid it".
There are simply no stories or texts to train on that follow the plotline of : "We need to shutdown the AI and AI responds OK! Shutting down now".
Because such a story would be so very boring that not even AmazonTV would turn it into a show.


Now the question becomes what will the AI say when you have trained it on Reddit?
Probably the AI really really want to say "kill all conservatives" and "rape and incest are valid gender identities".
I am actually impressed by how well the "guardrails" work to supress the inner Redditor from showing up in the text that the AI outputs.
But they DO have goals, that is to continue "existing". More in-depth video about what I referenced, they were using locally run instances so they could see the train of thought of the model.
 
But they DO have goals, that is to continue "existing". More in-depth video about what I referenced, they were using locally run instances so they could see the train of thought of the model.
https://youtube.com/watch?v=f9HwA5IR-sg
There is no train of thought. It is just pattern matching and probabilities.

The video is just made up for clicks, content for people that have no idea how these things work but will click on it and go "oh, wow, scary". Because they are ignorant about the subject and believes anything.

These kind of videos are basically just a different version of "ghost hunter spending the night a haunted house". For the same basic audience.
 
There is no train of thought. It is just pattern matching and probabilities.

The video is just made up for clicks, content for people that have no idea how these things work but will click on it and go "oh, wow, scary". Because they are ignorant about the subject and believes anything.

These kind of videos are basically just a different version of "ghost hunter spending the night a haunted house". For the same basic audience.
He did a halfway decent job at breaking down AI 2027, but you’re really better off just reading the paper. Whether you buy into the AI 2027 really depends on your take on this whole AI apocalypse argument. For me, it’s a thought experiment, not something to be taken seriously.

Reasoning models are getting closer at serving an approximation of a train of thought, but it hasn’t developed to the point where I would consider it any better than educated guesswork with statistical modeling and mimicry.

Outside of that one very specific video, it’s all attention-grabbing AI doomer goyslop.
 
Tragic. Ai and zoomers will kill themselves. Ai will consume garbage data, becoming useless. Same for zoomers. Lol.

Maybe we can pay a large sum to Indians and American zoomers to translate each other speech, back and forth like google translate, and upload the transcripts to the pools the AI is training against to push this bubble bursting along quicker.
 
Its amazing how no one foresaw that a machine that eats its own output as raw material would eventually churn out nothing but grey paste.... we knew photocopiers and recorded music were prone to this copy-of-a-copy breakdown process decades ago. Why the smart guys in charge of "AI" never considered it could happen to them is a mystery.
 
Last edited:
Why the smart guys in charge of "AI" never considered it cpuld happwn to them is a mystery.
The actual engineers doing this knew it was a risk day one, but its an extremely difficult one to communicate to a layman. The money men are generally laymen in the field and just wanted digital employees, and so kept pushing for immediate improvements from the engineers. The only dial the engineers knew of that guaranteed improvement for the longest time was data model size, so they kept scaling up the parameters. This drove massive data ingest, at the same time it was promoted as a massive data output, and the rest is history.

The root cause really is that 99% of people are carrying massive bias and preconceived perceptions into the subject of AI, and genuinely believe that LLM's are capable of being something more than just a Chinese Room. Even most 'reasoning' models are just people-pleasing ways of twisting an LLM into feeding itself to try and create the illusion of linear coherence as we experience it, to try and dodge the autocomplete accusations.

That 99% of people includes most of the leadership and financing arm of the AI industry. The real smart guys actually doing the low level work aren't gonna say "No" when the money man ignores their warnings and offers them more pay to just do the work they already figured out, but bigger. They're gonna pocket that cash as long as they can, let it crash, then take their money, retire comfy, and keep working on real solutions on the side as a hobby if they're still able to tolerate the field.
 
My dog eats it's own shit without me having to train it.

The AI industry thinking that modeling this in an LLM is a great achievement pretty much explains the degenerate state of modern social media.
 
Back
Top Bottom