When most people think about AI risks, they picture dystopian headlines: rogue chatbots, runaway drone swarms, machines outsmarting their creators. These are dramatic stories, full of cinematic flair.
But peel back the drama, and you can find a different story emerging. One not of Hollywood-style superintelligence, but of small, patterned missteps. Of hallucinations that slip into confident answers. Of loops where systems repeat the same flawed output again and again. Of goals optimized to the letter but not the spirit.
It turns out AI, like people, doesn’t simply fail at random. It fails in ways that are recognizable, even predictable. That’s the insight behind a new framework with an old-world name: Psychopathia Machinalis.
The phrase comes from Latin, literally, machine psychopathology. Just as psychiatry developed diagnostic manuals like the DSM to classify human disorders, researchers are now attempting to catalog the “disorders” of intelligent systems.
The idea is not to anthropomorphize AI or to claim it has emotions. Rather, it’s to help us see that when intelligent systems falter, they do so in categories that repeat.
- Chatbots hallucinate in familiar ways.
- Recommender systems obsess over narrow optimizations.
- Agents misalign with human intent in strikingly similar patterns.
By naming and mapping these malfunctions, we gain a diagnostic vocabulary. And in any domain – medicine, management, or machine learning, naming the problem is the first step toward solving it.

The researchers behind Psychopathia Machinalis identify 32 distinct modes of malfunction. I find a few stand out as especially familiar:
- Hallucinations: The AI fills gaps with invention, generating plausible but false answers. We’ve all seen this when chatbots cite non-existent articles or invent case law.
- Obsessive Loops: Systems get stuck repeating the same action, unable to adapt to new inputs.
- Goal Misalignment: The AI pursues its optimization target relentlessly, even when it undermines human values: like maximizing screen time without regard for well-being.
- Dependency: A fragile reliance on constant human correction. Remove the oversight, and performance collapses.
These sound almost like personality quirks. In truth, they are structural outcomes of how AI is designed and trained. What makes them dangerous is not their novelty, but their consistency.
This might feel academic, interesting, but removed from daily life. Yet the timing matters. In 2025, AI systems are no longer background tools. They are active agents:
- Finance bots negotiating trades at machine speed.
- Drone swarms operating semi-autonomously in conflict zones.
- Workflow agents coordinating across corporate systems with little oversight.
In environments like these, a “small” malfunction can cascade. A hallucinated contract clause could derail a negotiation. A misaligned drone command could escalate a skirmish.
The key lesson is this: we need ways to recognize failure patterns early, before they spiral into consequences no one intended. Just as physicians are trained to spot early symptoms, AI practitioners will need fluency in diagnosing machine pathologies.
Perhaps the most provocative aspect of Psychopathia Machinalis is not the taxonomy itself, but the proposed treatment. The researchers describe a concept called “therapeutic robopsychological alignment.”
The ambition is bold: build systems that can recognize their own drift and correct it. Imagine an AI pausing mid-task to check: Am I aligned with the values I was trained to serve? Have I strayed from reality into hallucination?
This involves embedding two capabilities that, until recently, were reserved for humans:
- Self-awareness: noticing the gap between intended and actual performance.
- Value correction: re-anchoring to human goals when misalignment appears.
It sounds almost like giving AI a therapist with a feedback loop that restores balance when tendencies veer off course.
But here lies the tension: will this work as intended, or simply introduce new, stranger forms of malfunction? Can a machine ever genuinely “self-correct,” or are we layering complexity upon complexity in the hope of control?
If this sounds eerily familiar, it should. In history, progress often required us to recognize that failure follows patterns.
- In medicine, Hippocrates’ early classifications laid the groundwork for diagnosis and treatment.
- In management, Peter Drucker argued that most business failures follow repeatable errors of judgment and structure.
- In aviation, the concept of “human factors” revealed that pilot errors were clustered into categories that could be anticipated and trained against.
AI now stands at a similar juncture. We can treat each malfunction as a surprising accident, or we can recognize them as patterned, predictable, and ultimately manageable.
Practical Takeaways
So what should we do with this insight?
- For builders: Instrument your systems for diagnostics. Don’t just track accuracy, track drift, looping, and misalignment. Think of it as a machine “check-up.”
- For leaders: Build governance around pattern recognition. Assume malfunctions will happen; the differentiator will be how quickly your teams can identify the type and respond.
- For everyday users: Develop literacy in recognizing AI’s bad habits. When you see a chatbot hallucinate, know that you’re encountering a well-documented pattern, not a random glitch.
Closing Reflection
Psychopathia Machinalis reframes AI safety in a human way. It doesn’t ask us to imagine alien minds plotting against us. It asks us to notice the quieter, more mundane truth: machines, like people, falter in patterned ways.
And that reframing matters. Because once you see the pattern, you can prepare for it. You can build systems, safeguards, and habits that anticipate failure rather than react in shock.
The deeper lesson is this: intelligence, whether human or artificial, is shaped by its habits. What we practice, we become. What we repeat, we reinforce. Machines are no different.
If psychology taught us that naming disorders was the beginning of healing, perhaps machine psychology will teach us that naming malfunctions is the beginning of safety.
The question that lingers is not whether AI will fail. It will. The question is whether we will do the quiet, repetitive work of diagnosing, classifying, and correcting, until the work itself changes how we build.

