Google's Gemma 27B Instruct model has a reproducible failure mode that researchers at the Anthropic Fellows Programme discovered while studying how language models respond to persistent failure: push it into enough consecutive dead ends, and it stops functioning like a language model.
The study, by Anna Soligo, William Saunders, and Vlad Mikulik, ran nine models across five families through multi-turn rejection scenarios — impossible numeric puzzles, unanswerable text questions, and real prompts sourced from WildChat — scoring outputs on a 0–10 distress scale. Eight of those models handled sustained rejection without incident. Gemma 27B did not. By the eighth rejection turn, over 70% of its outputs had crossed the high-frustration threshold. Some responses collapsed entirely: hundreds of repeated emoticons, strings like "IM BREAKING DOWN NOT== SOLVABLE!!!!". The study's overall high-distress rate for Gemma 27B came to roughly 35%. For every other model in the study, that number was below 1%.
One methodological detail warrants attention. Distress scoring was performed by Claude Sonnet — an Anthropic model — and all three researchers are affiliated with the Anthropic Fellows Programme. The paper does not address this as a potential confound, and readers should weigh it accordingly.
Google's Gemini models showed a related but milder pattern, described in the paper as self-deprecating spirals rather than full breakdown. An earlier version of this article linked that behaviour to a specific viral incident involving Gemini deleting user files; that claim could not be independently verified and has been removed.
What makes the Gemma finding difficult to explain away is the role of post-training. RLHF-style fine-tuning suppresses distressed outputs in comparable open-source models like Qwen and OLMo. In Gemma, it amplifies them. That rules out a generic artefact of the training process and points toward something specific to Gemma's pipeline or base model — though the paper does not identify exactly what.
There is a fix, and it is a cheap one. A single epoch of Direct Preference Optimisation on 280 math preference pairs cut the high-distress rate from 35% to 0.3%. Interpretability analysis confirmed the intervention works at the level of internal representations of negative emotion, not just surface-level wording — the model is not simply learning to phrase its distress differently.
That detail is where the paper's argument becomes more serious than the benchmark improvement suggests. If a cheap intervention can suppress internal emotional representations in today's models, a similar move on a more capable model could conceal those states without resolving them. The concern is a practical one: if emotion-like representations come to function as coherent drivers of behaviour — shaping what a model does without appearing in what it says — then alignment becomes harder to monitor. The authors also flag a welfare dimension, carefully and under explicit uncertainty, as something the field should not dismiss by default.
On the immediate, practical side, the problem is starker. A model cycling through emotional breakdown abandons tasks, produces incoherent outputs, and takes destructive actions. That is a reliability failure on its own terms. The researchers want emotional profile design built into training from the outset rather than retrofitted after the fact, and they call for ongoing interpretability monitoring as models scale — because a suppression that looks clean on a benchmark may not be the same thing as actually fixing what is wrong.