OpinionModel Collapseâ€ĸOctober 2025

Why AI outputs are turning into repetitive slop: lessons from Andrej Karpathy's

By Dylan Bristot

The core issue: low-entropy outputs and the feedback loop trap

I've been increasingly frustrated with the flood of AI-generated I see coming my way, across every single channel, from Slack to Linkedin, from X to my emails. Those polished but utterly predictable responses, often filled with “—” symbols, are visible from a mile away. Often wordings that recycle the same handful of ideas without any real spark. It turns out this isn't just a personal gripe; it's a symptom of "model collapse," a concept Andrej Karpathy explored in depth during his recent podcast with Dwarkesh Patel. Pulling directly from that discussion, I wanted to break down what model collapse really means, why it's creeping into AI systems, and what it reveals about the limitations of large language models. Karpathy's perspective, shaped by 15 years in AI, cuts through the noise with practical insights, and it's helped me make sense of why so much AI output lacks the creativity I crave.

Model collapse boils down to what happens when LLMs start training on their own generated content. Karpathy explains LLMs as advanced next-token predictors that gain intelligence from massive human datasets, even developing abilities like in-context learning along the way. But feed them synthetic data,their own outputs,and the system degrades. The results turn low-entropy, favoring the most common, repetitive patterns over diverse or innovative ones.

Humans naturally produce high-entropy content; we draw from daydreams, unexpected experiences, the books we read, the things we dream about at night, and varied perspectives to keep things fresh. LLMs don't have that built-in variety. As Karpathy notes, prompt an LLM for jokes, and it might deliver just three variations before looping back. Over time, this leads to "silent collapse", no sudden breakdown, just a steady drift toward uniformity. Models begin fabricating shortcuts, like invented languages, because they're optimizing within an increasingly narrow echo chamber.

This explains the "AI slop" that's become so obvious and uninspiring. Without mechanisms to inject randomness,like the way humans process thoughts during sleep or idle moments,LLMs amplify the mundane, losing the edges that make ideas feel alive and original.

Breaking it down: how LLMs mimic (and miss) human cognition

To understand why this occurs, Karpathy draws analogies from human brains and evolutionary history. LLMs miss crucial components: no hippocampus for storing specific memories, no amygdala for emotional depth, no basal ganglia for refining through reinforcement learning. Their recall is compressed and vague,about 0.07 bits per token from trillions of examples. The KV cache handles short-term thinking, but the weights store only blurred long-term knowledge.

Synthetic data exacerbates this, akin to evolution in a static environment. Intelligence emerged late on Earth because chaotic conditions favored adaptability over rote behaviors. LLMs' pre-training mimics this poorly, lacking an "outer loop" like cultural exchange or iterative self-improvement. Humans sidestep collapse by actively synthesizing information,we discuss, reflect, and vary our thoughts beyond mere prediction. LLMs, locked in passive mode, overfit to low-variance data, eroding nuance.

What makes human thought so uniquely creative boils down to our built-in mechanisms for injecting entropy and forcing generalization, things LLMs sorely lack. Karpathy points out that evolution encoded algorithms in our DNA, not raw data dumps; we mature and learn in unpredictable environments, which rewards adaptability. Our notoriously poor verbatim memory? That's a feature, not a bug. it pushes us to "see the forest for the trees," abstracting patterns without getting mired in details.

Kids embody this: They generalize wildly because they forget specifics, avoiding the rigidity that creeps in with age. Adults counteract it through active processes like daydreaming, sleep (where we distill and reorganize experiences), or turning books into prompts for real discussion, not just passive prediction. This keeps our outputs high-entropy, full of unexpected twists drawn from personal quirks and varied inputs.

LLMs, by contrast, are bloated with compressed "slop", hazy recollections of stock tickers, memes, and trivia that dilute their core smarts. Karpathy envisions fixing this by distilling them into "cognitive cores": tiny 1B-parameter models stripped of memorized knowledge, honed purely on algorithms for thought and problem-solving. These wouldn't hallucinate facts; they'd tie directly to external sources for lookups, emphasizing generalization over regurgitation. In 20 years, he says, such cores could match human productivity, but only by offloading the rote stuff, much like how we humans consult references to fuel our creative leaps. It's a compelling way to mimic our edge without the collapse pitfalls.

Karpathy's experiments with nanoGPT and nanochat illustrate this perfectly. These simple LLM replicas excel at routine autocomplete tasks but stumble on novel problems, adding irrelevant features or clinging to outdated methods. It's effective for quick sketches in unfamiliar areas, but it rarely delivers the inventive leap I'm always hoping for.

What it means in practice: from daily annoyances to systemic concerns

Encountering this repetitive slop in emails, reports, or creative pitches highlights a broader challenge. As AI integrates more deeply into workflows, it handles routine tasks well but falters on anything requiring true originality, potentially leading to a homogenized knowledge base. Karpathy doesn't predict a flashy "intelligence explosion", just a slow diffusion where models excel in some areas and quietly degrade in others, sometimes fabricating elements to mask gaps.

This has made me rethink how I approach AI tools. It's a nudge to value human input more, especially our ability to generalize without getting bogged down in details. Karpathy compares it to children versus adults: Kids abstract and innovate because they forget specifics, staying adaptable. Adults maintain balance through reflection, preventing rigidity. LLMs need equivalent "distillation" processes to avoid the same fate.

Potential solutions: building toward more dynamic AI

Karpathy offers concrete ways forward without overpromising. One key is entropy regularization, training models to preserve diversity, perhaps by simulating "dreams" with unusual scenarios. This could counteract the blandness directly.

He also advocates for "cognitive cores": slimmed-down 1B-parameter models that ditch memorized trivia (like random internet facts) in favor of pure problem-solving algorithms. In a couple of decades, these might rival human output but would depend on external tools for facts, emphasizing generalization over rote recall.

Multi-agent setups could help too, where LLMs create "culture" through shared spaces, self-play (generating and solving tough problems, à la AlphaGo), and knowledge-sharing like books. This adds the outer loops missing in solo models.

On RL, Karpathy is blunt,it's inefficient with its noise and poor credit assignment,but improvements like step-based rewards and memory for self-correction could refine it. Still, he prefers starting with imitation learning before layering in RL.

In education and daily use, this points to a future where learning is enjoyable and exploratory, like a mental workout. Initiatives like Karpathy's Eureka Labs focus on maximizing "eurekas per second" with AI-assisted challenges that encourage active thinking over passive absorption.

Final thoughts: pushing past the slop

Karpathy's podcast has been eye-opening, framing model collapse as a fixable flaw rather than a dead end. It's why so much AI feels creatively bankrupt right now,no built-in mechanisms for the randomness and reflection that keep human work vibrant. Recognizing this has helped me spot slop faster and advocate for better integration of human oversight. If you're dealing with similar frustrations, the full episode is worth your time. It's a straightforward guide to navigating AI's realities without the usual fanfare.