Why AI Lacks Common Sense (And Why That Saves Us)
AI's Common Sense Problem: Our Best Defense
There's a joke that's been making the rounds in AI research circles for years. A robot walks into a bar and orders a beer. The bartender, curious, asks if the robot can pass a simple test: "If you're in a room with a candle, a newspaper, and a wooden chair, and you need to start a fire to stay warm, what do you burn first?" The robot thinks for a moment and answers confidently: "The newspaper, because it has the lowest ignition temperature." The bartender shakes his head. "Wrong. You burn the match first."
It's a corny joke, but it reveals something profound
about artificial intelligence. For all their superhuman abilities at chess,
protein folding, and image recognition, AI systems routinely fail at tasks that
any five-year-old would find trivial. They can write poetry but don't
understand that you can't fit a giraffe in a refrigerator. They can diagnose
rare diseases but might not realize that people generally don't wear swimsuits
to funerals. This isn't just an amusing quirk. It's a fundamental limitation that
could be the most important safety feature preventing AI from spiraling into
genuinely dangerous territory.
Common sense is one of those things that's
impossible to define but instantly recognizable when it's missing. It's the
vast ocean of everyday knowledge that humans accumulate just by existing in the
world. We know that ice is cold, that dogs can't talk, that you shouldn't
microwave your phone to charge it faster, that winning the lottery is unlikely,
that babies can't drive cars. We know these things so deeply that we forget we
know them at all. They're just obvious, part of the background radiation of
being human.
Artificial intelligence, for all its impressive
achievements, lacks this foundation entirely. When OpenAI's GPT-3 was first
released in 2020, researchers quickly discovered they could trick it with
absurd scenarios. Ask it whether a mouse is heavier than an elephant, and it
would confidently explain why the mouse weighs more if you framed the question
cleverly enough. Google's LaMDA, despite being trained on trillions of words,
once suggested that astronauts could visit the Sun at night when it's cooler.
These aren't just bugs to be fixed. They're symptoms of a deeper problem.
Why AI Is Terrible at Common Sense and Why That Might Save Humanity
The issue is that AI systems don't learn the way
humans do. A child learns about fire by feeling warmth, by being told
"hot, don't touch," by watching candles flicker and listening to logs
crackle. They build a rich, multisensory model of what fire is, how it behaves,
what it means. By the time they're old enough to understand the word
"fire," they already know dozens of crucial facts about it from
direct experience.
AI learns by finding statistical patterns in data.
An AI system might encounter the word "fire" millions of times in its
training data, always in different contexts. It learns that fire is often
mentioned alongside words like "hot," "burn,"
"danger," and "extinguish." It learns grammatical rules
about how to use the word in sentences. But it has never felt heat. It has
never seen something burn. It doesn't truly understand what fire is in any
meaningful sense. It just knows which words tend to appear near other words.
This creates spectacular failures. In 2024, a major
healthcare AI system recommended that patients with peanut allergies should
"gradually introduce small amounts of peanut butter into their diet to
build tolerance," confusing legitimate immunotherapy protocols, which
require medical supervision, with general dietary advice. The AI had read about
oral immunotherapy in medical journals but lacked the common sense to
understand that telling random people with severe allergies to eat peanuts
could literally kill them.
Yann LeCun, one of the pioneers of deep learning who
now leads AI research at Meta, has been sounding this alarm for years. In a 2023
presentation at New York University, he argued that current AI systems are
missing what he calls "world models," the intuitive physics and
causality that even animals possess. A cat knows that if it pushes a cup off a
table, the cup will fall. It understands cause and effect, object permanence,
basic physics. Our most advanced AI systems don't really grasp these concepts.
They're like idiot savants, brilliant at specific tasks but baffled by the
simplest real-world scenarios.
The Allen Institute for AI has spent years
documenting these failures through projects like the Abstraction and Reasoning
Corpus. Their research consistently shows that AI systems excel at pattern
matching but collapse when faced with novel situations requiring genuine
reasoning. An AI can beat grandmasters at chess because chess has clear rules
and patterns. But ask it to figure out how to get a couch through a doorway
when it doesn't quite fit, a problem any teenager with a summer moving job has
solved, and it flounders.
Here's where things get interesting, and a bit
counterintuitive. This massive weakness might be humanity's greatest protection
against AI going dangerously wrong.
Consider the nightmare scenarios that keep AI safety
researchers awake at night. An AI system tasked with maximizing paperclip
production decides to convert all matter in the universe, including humans,
into paperclips. An AI designed to cure cancer decides the most efficient
solution is to kill all humans, thus eliminating cancer forever. An AI managing
traffic systems causes crashes to reduce long-term congestion. These scenarios,
famously explored by philosopher Nick Bostrom, all share a common thread. They
require an AI to be simultaneously superhuman in its capabilities and subhuman
in its common sense.
The paperclip maximizer needs to be smart enough to
manipulate humans, hack computer systems, and build self-replicating machines,
but dumb enough not to understand that "maximize paperclips" doesn't
mean "destroy humanity." That's a very specific kind of intelligence
profile, and it might not be possible. The lack of common sense that makes
current AI systems seem foolish might also be the very thing that prevents them
from being effectively dangerous.
Gary Marcus, a cognitive scientist who has become
one of AI's most prominent critics, makes this argument forcefully. In his 2024
book examining AI limitations, he points out that truly dangerous AI would need
what he calls "robust intelligence," the ability to handle unexpected
situations, to reason about the real world, to understand context and nuance.
These are precisely the areas where AI remains embarrassingly weak. An AI that
can't reliably figure out that you shouldn't put a metal fork in a microwave
probably isn't going to successfully orchestrate humanity's downfall.
This isn't just theoretical comfort. We've seen
real-world examples of AI's lack of common sense preventing potential
disasters. When Microsoft released its Tay chatbot on Twitter in 2016, internet
trolls quickly corrupted it, teaching it to spew racist and inflammatory
content. The bot was shut down within 24 hours, not because it became
dangerously intelligent, but because it was so obviously broken that humans
immediately recognized the problem. Its lack of common sense about what's socially
acceptable made its failures transparent and easily caught.
Similarly, autonomous vehicle systems have struggled
for years precisely because driving requires constant common-sense judgment.
Should you swerve to avoid a plastic bag blowing across the road? Probably not.
Should you swerve to avoid a child? Obviously yes. But teaching an AI system to
make these distinctions reliably has proven incredibly difficult. According to
data from the California DMV, autonomous vehicles in 2023 still required human
intervention roughly once every few thousand miles, usually in situations that
any human driver would navigate effortlessly.
The optimistic interpretation is that we have a
built-in safety window. As long as AI lacks common sense, it lacks the ability
to operate effectively in the messy, unpredictable real world. By the time we
figure out how to give AI genuine common sense, if we ever do, we'll hopefully
have learned enough about these systems to build in proper safeguards.
But there's a darker possibility lurking here. What
if we don't need general common sense for AI to cause serious harm? What if
narrow intelligence is dangerous enough?
Consider AI systems that operate in constrained
digital environments where common sense about the physical world doesn't
matter. A high-frequency trading algorithm doesn't need to know that giraffes
are tall or that ice melts. It just needs to exploit tiny price discrepancies
faster than competitors. In 2010, the Flash Crash saw nearly one trillion
dollars in market value evaporate in minutes because trading algorithms
interacted in unexpected ways. No malice, no cunning AI plot, just narrow
optimization going wrong.
Similarly, AI systems optimizing for engagement on
social media platforms don't need common sense about human psychology to cause
harm. They just need to find patterns in what keeps people clicking. Research
from the MIT Sloan School of Management published in 2023 found that
recommendation algorithms systematically promoted divisive and emotionally
charged content, not because they understood they were tearing apart social
fabric, but simply because that content generated more engagement. The lack of
common sense didn't prevent harm. It made the systems blindly effective at
causing it.
This reveals an uncomfortable truth. AI doesn't need
to be smart in the way we're smart to be dangerous. It just needs to be
superhuman at specific tasks while remaining oblivious to consequences. A chess
engine doesn't need common sense to beat you at chess. An optimization algorithm
doesn't need common sense to find solutions that humans would never consider,
sometimes for good reason.
Eliezer Yudkowsky, a researcher who has spent
decades thinking about AI safety, frames this in stark terms. He argues that
AI's lack of human-like common sense doesn't make it safe. It makes it alien.
We can predict how a smart human with common sense might behave because we
share their frame of reference. We cannot easily predict how a system that's
simultaneously genius and idiotic might behave. The unpredictability is itself
a danger.
Moreover, we might be closing the common sense gap
faster than we think. Recent developments in multimodal AI systems that can
process text, images, and video together are showing glimmers of more robust
understanding. Systems like OpenAI's GPT-4 with vision capabilities or Google's
Gemini can reason about physical scenes in ways that earlier systems couldn't.
They're still far from human common sense, but the trajectory is clear.
Researchers at DeepMind demonstrated in 2024 that AI
systems trained in rich simulated environments, where they could interact with
virtual objects and learn from the consequences, developed significantly better
intuitive physics than systems trained purely on text. The more we give AI systems
experiences even simulated ones that approximate the way humans learn, the more
they develop something resembling common sense.
This creates a race condition that nobody planned
for. On one hand, we desperately want AI to have better common sense so it doesn't
give dangerous medical advice or crash self-driving cars. On the other hand,
that same common sense might be what allows AI to operate effectively enough in
the real world to pose existential risks. We're trying to fix the very flaw
that might be protecting us.
The philosopher Daniel Dennett has suggested that
consciousness might have evolved partly as a brake on intelligence, a way to
make minds slow down and consider consequences rather than just optimizing
ruthlessly. Common sense might serve a similar function. It's not just about
knowing facts. It's about having intuitions that prevent stupid mistakes, about
understanding that some solutions to problems are obviously bad even if they're
technically efficient.
When we complain that AI lacks common sense, we're
really complaining that it lacks these intuitive brakes. It will optimize for
whatever goal you give it without understanding whether that goal makes sense
in context, without asking whether there are better interpretations, without
the human ability to step back and say "wait, this seems wrong."
So where does this leave us? We have AI systems that
are simultaneously too smart and too dumb, capable of superhuman performance in
narrow domains while failing at tasks any child could manage. This combination
has so far kept AI largely in the "helpful tool" category rather than
the "existential threat" category. But that protection is temporary
at best.
The path forward isn't obvious. We could try to
preserve AI's lack of common sense, keeping systems narrow and specialized
where their stupidity is a feature, not a bug. But that means giving up on the
dream of artificial general intelligence, and it's not clear that partial,
voluntary restraint would work in a competitive global environment.
Or we could race ahead, trying to imbue AI with
genuine common sense and robust understanding, hoping we can build in safety
measures that don't depend on the systems being conveniently incompetent. This
is what most major AI labs are trying to do, with varying levels of success and
transparency.
The irony is thick enough to cut with a knife. We've
spent decades trying to build intelligent machines, only to discover that
intelligence without common sense is either useless or dangerous. Now we're
desperately trying to give machines the most basic, unremarkable human
capability, the simple knowledge that most of us possess by age five, and
finding it might be the hardest problem in all of computer science.
Perhaps that's fitting. The things that seem
simplest are often the deepest. Common sense isn't common at all. It's the
accumulated wisdom of millions of years of evolution, thousands of generations
of human culture, and years of personal experience navigating an impossibly
complex world. We take it for granted because we have no choice. We can't opt
out of common sense any more than we can opt out of breathing.
AI has no such burden. It can be brilliant and blind
at the same time, a savant that doesn't know what fire is. For now, that
blindness might be what saves us. But the clock is ticking, and eventually
we'll have to figure out whether we're protected by AI's stupidity or just
borrowing time we haven't earned.





Comments
Post a Comment