Blog
Notes on engineering, design, and what I'm learning while building.
Notes on engineering, design, and what I'm learning while building.
Explore how AI hallucinations have been creatively leveraged to embark on new discoveries, and whether the quest for perfect accuracy might eliminate the accidents that occasionally lead to breakthroughs.
Visual metaphor for AI “hallucinations”: generative leaps that can mislead or inspire.
In 2024, Dr. David Baker received the Nobel Prize in Chemistry for revolutionary work in protein design. What the Nobel committee didn't emphasize—and what they deliberately avoided calling by name—was that Baker's breakthrough discoveries came from AI systems that were, by technical definition, "hallucinating."
Using Google DeepMind's AlphaFold, Baker's team generated over ten million completely novel protein structures that had never existed in nature. The AI wasn't just predicting existing proteins—it was inventing entirely new ones, creating biological architectures that evolution had never explored. These "hallucinated" proteins led to roughly 100 patents, the founding of over 20 biotech companies, and ultimately, a Nobel Prize.
This story embodies one of the most fascinating paradoxes in artificial intelligence today: the same mechanism that causes AI systems to confidently state falsehoods as facts—what we call "hallucination"—may also be the source of their most creative and innovative capabilities.
But this raises profound questions. If hallucinations can lead to Nobel Prize-winning discoveries, are they really errors? And if we eliminate AI's tendency to hallucinate, might we also eliminate its capacity for creativity and innovation?
The term "AI hallucination" itself is controversial. In humans, hallucinations are false sensory experiences—seeing things that aren't there, hearing voices that don't exist. AI systems don't have senses, so they can't hallucinate in the traditional sense. What they do is more akin to what psychologists call "confabulation"—the creation of false but coherent narratives to fill gaps in knowledge.
When ChatGPT invents a non-existent academic paper complete with convincing author names, publication dates, and abstracts, it's not malfunctioning—it's doing exactly what it was trained to do: predict the next most likely word based on patterns in its training data. The problem is that "most likely" doesn't always mean "true."
Recent research reveals that even the most advanced AI models hallucinate with surprising frequency. Google's Gemini-2.0-Flash, currently the most reliable large language model, still generates false information 0.7% of the time. At the other end of the spectrum, some models hallucinate in nearly one out of every three responses—a rate of 29.9%.
What's particularly fascinating is that more advanced AI systems—those capable of sophisticated reasoning—often hallucinate more than their predecessors. OpenAI's o1 reasoning model hallucinated 16% of the time when summarizing information about people, while their newer o3 and o4-mini models hallucinated 33% and 48% of the time, respectively.
This isn't to say that hallucinations are intentional features to be celebrated—the vast majority are simply errors that provide no value and can cause real harm. Rather, the evidence suggests that the underlying mechanisms that occasionally produce useful creative outputs are inseparable from those that generate useless or harmful falsehoods. It's a mathematical inevitability in current AI architectures: you can't have one without the other.
To understand why AI hallucinations might be valuable, we need to examine how human memory and creativity actually work—because humans "hallucinate" far more than we'd like to admit.
Human memory isn't a recording device; it's a reconstruction system. Every time we recall an event, we're not playing back a stored file but actively rebuilding the memory from fragments, filling in gaps with plausible details, updating it with new information, and coloring it with current emotions and biases.
Recent groundbreaking research from MIT's Media Lab demonstrates just how malleable human memory really is. In controlled experiments, researchers showed that AI-edited images and videos can implant false memories in human subjects with startling effectiveness. Participants exposed to AI-generated videos of edited images were 2.05 times more likely to recall events that never happened, and they reported higher confidence in these false memories than in real ones.
The study's implications are profound: our brains are essentially biological hallucination machines, constantly creating convincing narratives that may have little resemblance to reality. As Harvard memory researcher Daniel Schacter explains, "Memory is more than just the attic of our minds—it's a foundational part of how we interpret and imagine our futures."
This unreliability isn't a design flaw—it's a feature. Human creativity depends on our ability to recombine memories in novel ways, to see connections that don't exist, and to imagine scenarios that never were. The same neural processes that enable us to envision the future, compose music, or solve problems creatively are the ones that make our memories unreliable.
False memories serve an essential creative function. They allow us to:
Research in cognitive science shows that human creativity isn't just about retrieving information—it's about recombining and expanding existing knowledge, often through processes that closely resemble the "hallucinations" of AI systems.
Some of history's greatest scientific breakthroughs began as what could be called "hallucinations"—ideas that seemed to contradict established facts but ultimately revealed deeper truths.
Consider heliocentrism: for centuries, Copernicus's claim that Earth orbited the Sun was dismissed as a "factual hallucination" that contradicted observable evidence. People could clearly see the Sun moving across the sky and feel the Earth standing still beneath their feet. Yet this "false" idea revolutionized our understanding of the cosmos.
Similarly, Alexander Fleming's discovery of penicillin came from what could be considered a "faithfulness hallucination"—noticing something that shouldn't have been there (mold killing bacteria in a contaminated petri dish) and imagining its potential rather than dismissing it as experimental error.
These examples illustrate a crucial point: innovation often requires the ability to see beyond current facts, to imagine possibilities that don't yet exist, and to trust in patterns that may not be immediately verifiable.
Today's AI-powered discoveries follow similar patterns. Beyond Baker's protein work, researchers across multiple fields are harnessing AI hallucinations for breakthrough discoveries:
Medical Innovation: At California Institute of Technology, researchers used AI hallucinations to design a revolutionary catheter with sawtooth-like spikes that prevent bacterial contamination—a design that human engineers likely would never have conceived.
Drug Discovery: AI systems are generating novel molecular structures that don't exist in any chemical database. While many are impossible to synthesize, some represent entirely new classes of therapeutic compounds. This approach has dramatically accelerated drug discovery timelines from years to months.
Materials Science: AI hallucinations are creating new material compositions with properties that violate conventional wisdom, leading to innovations in everything from solar panels to superconductors.
Weather Prediction: Meteorologists use AI to generate thousands of subtle forecast variations, helping identify unexpected factors that influence extreme weather events—patterns that human forecasters might never consider.
The potential for AI hallucinations to drive innovation doesn't negate their serious risks. Recent legal cases highlight the dangers of unchecked AI creativity:
Legal Disasters: A Stanford Law School study found that when asked about legal precedents, AI models collectively invented over 120 non-existent court cases, complete with convincing case names and legal reasoning. One lawyer, relying on ChatGPT for research, submitted legal briefs citing three entirely fictional cases.
Medical Misinformation: Even the best AI models hallucinate potentially harmful medical information 2.3% of the time. In a field where accuracy can mean the difference between life and death, this error rate is unacceptable.
Corporate Consequences: Air Canada was forced to pay damages when its customer service chatbot hallucinated a non-existent bereavement fare policy, confidently telling a customer he could retroactively claim a discount that didn't exist.
Recent research from Wharton and published in Science Advances reveals a troubling paradox: while AI enhances individual creativity, it reduces collective diversity. When writers use AI assistance, their individual stories become more creative and better written. However, AI-assisted stories are significantly more similar to each other than human-only stories.
This creates what researchers call a "social dilemma": individuals benefit from AI assistance, but society loses the diversity of ideas that drives innovation. As AI systems become more widely adopted, we risk creating a world where everyone becomes individually more creative but collectively less original.
The implications are profound. If everyone uses the same AI tools, drawing from the same training data, we might see a homogenization of creative output—a world where individual creativity peaks but cultural innovation stagnates.
The hallucination debate forces us to confront fundamental questions about the nature of understanding, creativity, and truth itself. If an AI system generates a false statement that leads to a true discovery, was the original statement "wrong"? If a hallucination inspires a breakthrough, is it still an error?
Consider how some researchers now prefer the term "confabulation" over "hallucination" when describing AI errors. Confabulation, borrowed from neurology, describes the brain's tendency to fill gaps in memory with plausible but false information. It's not lying—it's creative gap-filling.
Journalist Benj Edwards suggests this terminology captures something crucial: these aren't random errors but "creative gap-filling" processes that may be essential to how both artificial and biological intelligence work.
Critics argue that terms like "hallucination" and "creativity" inappropriately anthropomorphize AI systems. They contend that these machines don't truly understand anything—they're simply very sophisticated pattern-matching systems that occasionally generate useful false patterns.
But this critique raises deeper questions: How different is human creativity from sophisticated pattern matching? When we have a breakthrough insight, are we doing something fundamentally different from recombining existing patterns in novel ways?
Some philosophers, following Harry Frankfurt's analysis, argue that AI output is more like "bullshit" than hallucination—generated without regard for truth or falsehood, accidentally accurate when correct and accidentally false when wrong. This view suggests that AI systems are fundamentally indifferent to truth, focused only on generating plausible-sounding responses.
The challenge isn't eliminating AI hallucinations—research suggests they're mathematically inevitable in current AI architectures—but learning to work with them productively while minimizing harm. This requires developing new frameworks for when we can tolerate uncertainty and when we must demand accuracy.
High-Stakes Contexts: In healthcare, legal proceedings, and financial advice, accuracy must be paramount. Here, we need robust verification systems, human oversight, and AI systems specifically tuned to minimize false information.
Creative Contexts: In art, research, brainstorming, and innovation, some level of hallucination may be beneficial or even necessary. The key is transparency—users should know when they're working with AI in "creative mode" versus "factual mode."
Hybrid Approaches: New techniques like Retrieval-Augmented Generation (RAG) allow AI systems to ground their responses in verified information while still maintaining creative capabilities. These approaches reduce hallucinations in factual contexts while preserving innovation potential.
Researchers are developing sophisticated approaches to manage the creativity-accuracy trade-off:
Self-Consistency Checking: Google's Gemini models use techniques where the AI compares multiple possible answers and selects the most coherent one, reducing hallucination rates by up to 65%.
Chain-of-Thought Prompting: Breaking complex tasks into smaller steps can reduce hallucinations while maintaining creative problem-solving capabilities.
Temperature Control: AI models can be adjusted to be more conservative (lower "temperature") for factual tasks or more creative (higher temperature) for exploratory work.
Meta-Awareness: Some AI systems are being trained to recognize when they're uncertain, with prompts like "Are you hallucinating right now?" reducing error rates by 17% in subsequent responses.
As AI becomes more integrated into creative and professional workflows, we need new forms of AI literacy. Users must understand:
Perhaps most intriguingly, researchers are exploring whether controlled AI-generated false memories could have therapeutic benefits. MIT studies suggest that AI-edited images could help patients with PTSD by allowing them to "remember" traumatic events differently, reducing their emotional impact under professional supervision.
This isn't about replacing real memories with false ones, but about harnessing the natural malleability of human memory for healing. Just as our brains naturally edit and reframe memories over time, AI could accelerate this process in beneficial directions.
Self-Esteem Enhancement: AI could subtly enhance images of past achievements, helping people build confidence by strengthening positive memories.
Anxiety Reduction: For people with social anxiety, AI could help create more positive associations with challenging situations by editing memories of past successes.
Skill Development: Athletes and performers could use AI-enhanced visualizations of perfect performances to build muscle memory and confidence.
The paradox of AI hallucination reveals something profound about the nature of intelligence itself—both artificial and human. The same underlying mechanisms that cause AI systems to confidently state falsehoods as facts—what we call 'hallucination'—also appear to enable their most creative and innovative capabilities.
Rather than viewing this as a problem to solve, we might see it as a feature to manage. The goal, I think, isn't to create perfectly accurate AI systems—such systems might lose the capacity for the creative leaps that occasionally lead to Nobel Prizes and revolutionary discoveries. Instead, we need to become more sophisticated about when we value accuracy and when we're willing to tolerate uncertainty in exchange for creative potential.
AI hallucinations represent a new form of creative commons—a vast space of generated possibilities, some true, some false, but all potentially inspiring. Like any commons, it requires careful stewardship. We need:
As AI systems become more creative, we're forced to confront uncomfortable questions about human exceptionalism. If an AI can write a moving poem, compose beautiful music, or design revolutionary proteins, what makes human creativity special?
Perhaps the answer lies not in the outputs but in the experience. Human creativity involves consciousness, emotion, personal meaning, and the lived experience of discovery. AI creativity, no matter how impressive, remains a statistical process without subjective experience.
But this doesn't make AI creativity less valuable—it makes it different. The future likely lies not in human versus AI creativity, but in new forms of collaboration where human insight guides AI exploration, and AI capabilities extend human imagination.
The paradox of AI hallucination teaches us something important about living with uncertainty in an age of artificial intelligence. Perfect accuracy may be neither achievable nor desirable. Instead, we need to develop new forms of wisdom: knowing when to trust, when to verify, when to explore, and when to constrain.
The most profound implications of AI hallucination may be philosophical rather than practical. They force us to confront the constructed nature of all knowledge, the creative power of false beliefs, and the fine line between delusion and insight that has always characterized human progress.
In embracing productive hallucinations while guarding against harmful ones, we're not just managing AI systems—we're learning to navigate the fundamental tensions between truth and creativity, accuracy and innovation, that define intelligence itself.
As we stand at the threshold of an age where artificial minds can dream up new realities, the question isn't whether AI will hallucinate—it's whether we'll be wise enough to distinguish between the hallucinations that mislead and those that illuminate the path forward.
Perhaps the greatest hallucination would be believing we can have perfect accuracy without sacrificing the creative uncertainty that makes discovery possible. The future belongs not to those who eliminate AI's capacity for productive falsehoods, but to those who learn to choreograph the dance between truth and imagination, reality and possibility, facts and dreams.
AI Hallucination Report 2025: Which AI Hallucinates the Most?
Comprehensive industry-wide statistics showing current hallucination rates across major AI models, from Google's Gemini-2.0-Flash (0.7%) to higher rates in other systems.
New Sources of Inaccuracy? A Conceptual Framework for Studying AI Hallucinations
Harvard Kennedy School analysis of AI hallucinations as a new form of misinformation with unique structural characteristics.
Beyond Hallucination: Generative AI as a Catalyst for Human Creativity and Cognitive Evolution
2025 academic paper reframing AI hallucinations as windows into artificial and human cognition rather than mere errors.
Hallucination (Artificial Intelligence) - Wikipedia
Comprehensive overview including David Baker's Nobel Prize-winning protein work and other scientific applications of AI hallucinations.
In Praise of AI Hallucinations: How Creativity Emerges from Uncertainty
Analysis of how the 2024 Nobel Prize in Chemistry was enabled by AI hallucinations and the broader implications for scientific discovery.
Harnessing Hallucinations to Make AI More Creative
Psychology Today examination of recent research demonstrating how AI hallucinations drive advances in drug discovery and materials science.
Synthetic Human Memories: AI-Edited Images and Videos Can Implant False Memories
MIT Media Lab groundbreaking study showing AI-edited visuals increase false recollections by 2.05x compared to controls.
AI-Implanted False Memories – MIT Media Lab Project
Comprehensive research project examining how AI systems can induce and amplify false memories across multiple modalities.
Extreme Amnesia Cases, AI, and Our Imagined Futures: In Conversation with Harvard Memory Researcher
National Geographic interview with Harvard's Daniel Schacter on how memory works, its relationship to imagination, and parallels with AI systems.
Confabulation Machines: Could AI Be Used to Create False Memories?
Analysis of the connection between AI hallucinations and human confabulation, with implications for memory manipulation.
Generative AI Enhances Individual Creativity but Reduces Collective Diversity
Science Advances study showing the social dilemma where AI improves individual creativity while reducing collective novelty.
Does AI Limit Our Creativity?
Wharton research on how ChatGPT boosts individual idea quality while reducing team diversity, creating trade-offs for innovation.
AI Can Catalyze and Inhibit Your Creativity
World Economic Forum analysis of the dual effects of AI on human creativity and productivity.
When AI Dreams: The Creative Potential of Hallucinations
Industry perspective on embracing AI hallucinations as sources of creativity and innovation while managing risks.
Not All Hallucinations Are Bad: The Constraints and Benefits of Generative AI
NTT DATA analysis of how to harness AI hallucinations productively while implementing responsible frameworks.
What Are AI Hallucinations?
IBM's comprehensive guide to understanding AI hallucinations, including positive applications in gaming, art, and data visualization.
Will Generative AI Ever Fix Its Hallucination Problem?
American Bar Association analysis of hallucinations in legal contexts, including the Michael Cohen case and judicial responses.
When AI Gets It Wrong: Addressing AI Hallucinations and Bias
MIT Sloan comprehensive guide to understanding and mitigating AI hallucinations in educational and professional contexts.
Hallucinations in AI: Fatality or Opportunity?
Human Technology Foundation exploration of hallucinations as creative potential, drawing parallels to historical scientific breakthroughs.
Hallucination vs. Confabulation: Rethinking AI Error Terminology
Analysis of why "confabulation" may be more accurate than "hallucination" for describing AI errors, with implications for understanding the phenomenon.
AI and Memory
Cambridge academic examination of how AI changes what memory is and does, creating new forms of human-machine memory interaction.