TL
TL
Hachette Book Group supports the right to free expression and the value of
copyright. The purpose of copyright is to encourage writers and artists to produce
the creative works that enrich our culture.
Hachette Books
Hachette Book Group
1290 Avenue of the Americas
New York, NY 10104
HachetteBooks.com
Twitter.com/HachetteBooks
Instagram.com/HachetteBooks
The Hachette Speakers Bureau provides a wide range of authors for speaking
events. To find out more, go to www.hachettespeakersbureau.com or call (866)
376-6591.
The publisher is not responsible for websites (or their content) that are not owned
by the publisher.
E3-20211201-JV-NF-ORI
CONTENTS
Cover
Title Page
Copyright
Dedication
Epigraph
Acknowledgments
Discover More
Notes
For Julie,
who saved me
WHENEVER OUR FUTURE on this planet looks bleak, we can’t help but think
about other planets. We’ve spilled over our place in nature, and we can’t seem to
get along well enough to agree on our shared salvation. Let’s go somewhere else
and start over. How far away is the next habitable planet, anyway?
Interplanetary scientists despise the notion of translating light-years to a speed
that you and I can grok as a number, the way we put a number to the speed of our
own cars, so it always falls to journalists like me to thumbnail a best guess as to
just how long it might take to carry us that same distance. Bear with me, please.
Our current rocket technology can propel a ship through space at speeds of
roughly 20,000 miles per hour. By the standards you and I are used to, that’s
incredibly fast. In our own atmosphere the friction against the surrounding air at
that speed would melt through any material we’ve invented and would incinerate
you and me before we could even strike up a conversation about where we were
going and what we wanted to build when we got there. But let’s use that speed as
a benchmark, because in space that speed is a terribly slow rate to cross the
enormous distances between planets.
Mars, whose orbit is directly adjacent to our own, is the most survivable other
planet in our solar system. But that’s not saying much. Sure, other planets are
more horrible. Piercing the thirty-mile layer of clouds that surrounds Jupiter, our
rocket’s engines would begin to either choke or over-fire in its flammable
nightmare of hydrogen and helium, and then would die completely as they hit the
liquid form of the same stuff roughly 13,000 miles past the cloud cover that keeps
us from knowing anything about what’s beneath that poisonous ocean. That crew
would drown (or maybe drown and burn, unheard of on Earth) without ever
making it out of the ship.
Mars is comparatively pleasant. For one thing, there’s stable footing. And a
nice day on Mars might actually feel nice. With the sun high in the sky, you’d
enjoy temperatures as high as 68° F, a clear August afternoon in San Francisco or
Johannesburg. But if you happened to exit the ship at one of the poles, at night,
the temperature could be less than −200° F, cold enough to not only instantly kill
you, but also almost instantly freeze your limbs into brittle branches that would
shatter under a hammer blow. And let’s not forget that even in the balmiest
regions of the planet, there’s nothing to breathe, so you’re not getting far on even
the most pleasant day. You’d bound perhaps fifty yards in the low gravity before
you could no longer hold your breath, then hypoxia would disorient you, and you
couldn’t make it back to the oxygen of the ship. You’d wind up unconscious and
twitching, your heart would stop, and the red dust would slowly settle around
your corpse.
That’s why scientists and journalists alike are so excited about exoplanets, the
term for planets beyond our solar system that seem to offer the possibility of a
livable atmosphere and surface. Humanity has been treated in the last few years to
a steady stream of optimistic fantasy destinations emanating from the now-
defunct Kepler space telescope. Kepler essentially squinted out into deep space to
see how light from distant stars bent around intervening planets too far away to
image in any detail. Depending on how the light goes around a planet on its way
to us, astrophysicists can calculate not only the size of that planet, but how far it
is from the source of light, meaning we can determine whether the relation
between the planet’s size and the distance to its star possibly indicates that planet
might host some sort of atmosphere.
The Kepler mission measured the light from roughly 150,000 stars and found
several hundred planets whose ratio of size and star-distance makes them
candidates—just candidates, but a real possibility—for human respiration and
occupancy. Land, walk around, build a cabin out of whatever materials exist, just
imagine! And when we consider the vast distances of space, the closest
exoplanets are, in fact, relatively close.
But before we pop the champagne and pour our savings into SpaceX, let’s
think about what it takes to get to another planet. A trip to Mars, for instance, is
comparatively brief. Depending on where it and Earth are in their orbits, the
journey could be made in between three hundred and four hundred days. But
humans have never traversed open space for that amount of time. The journey to
the moon currently takes about seventy-two hours, and astrophysicists and
medical experts quietly point out in private conversation that it’s a miracle none
of the two dozen people who went to the moon died during the trip. A trip to
Mars would involve exposing the crew to the dangers of deep space for roughly a
full year. And those dangers go on and on. Deadly amounts of radiation permeate
everything in the deep blackness between planets. Space is full of dirt and grit
that could disable the ship. (A whole field of astrophysics studies the interstellar
medium and has shown that if you held a gloved white hand out the window of
the ship as one does on the highway, it would come back blackened by, well,
whatever it is that’s out there.) Also consider that if a mishap killed the ship and
the crew, the event would be torturously broadcast, on time delay, to the whole of
Earth, presumably killing off our species’ desire to travel to Mars in the process.
And even if all goes well for the crew, that’s a long time confined together in a
space no bigger than a vacation rental, as all of us who spent the pandemic year
locked in with family, roommates, or alone know too well. In fact, before the
coronavirus made test cases of us all, psychologists and logisticians who worried
about a Martian crew driving each other nuts spent time observing actual
astronauts confined in these sorts of tiny spaces for the duration, either of a
simulated trip to Mars or a stay on the planet. And it hasn’t gone well. On almost
every sardine-style simulation someone has suffered serious injury or illness. A
simulated Martian habitat on a military base on the island of Hawaii has seen a
half-dozen such missions over the years, including one where a crew member had
to withdraw for medical reasons (the mission’s organizers haven’t publicly
revealed what it was). Seeking to learn from the experience, the crew pretended to
treat their missing member as dead and enacted setting a fake body out on the
simulated Martian tundra, where it would be perfectly preserved for a journey
back to Earth for burial. In the final Hawaiian mission, the simulation was
compromised when one of the crew was electrified by live wiring, and earthly
paramedics had to come inside and drive the crew member away in an ambulance.
But putting aside the physical danger of living isolated on Mars, the missions
have revealed that… people get weird. “You can select a crew all you want, get
the right fit and mix, but there’s too many variables when it comes to human
beings,” a psychologist for the mission told the Atlantic. “It’s just really hard to
predict how we’re going to perform in all situations.”
That’s just Mars. It’s next door to us, cosmically speaking. Now imagine how
weird we’d become trying to reach the nearest exoplanet.
Let’s imagine we’re standing together on the launch pad at NASA’s Cape
Canaveral facility near Orlando, and staring up at the stars together. As I write
this, the last constellation above the horizon is Centaurus. The centaur’s front
hoof is a bright star. In fact, it’s three stars—a pair called Alpha Centauri A and
B, and, dimmest of the trio, Proxima Centauri. Here, look through this telescope.
See? You can tell them apart. But what we can’t see is that there is, in fact, a
planet circling the faint light of Proxima Centauri. Man, I wish we could see it.
Because that planet, Proxima Centauri b, is the nearest known exoplanet to Earth.
We have no idea what life would be like on Proxima Centauri b, or what the
place even looks like. There may be many reasons that it just won’t work for
human habitation. It’s possible that stellar winds may push so much radiation
across its surface that we’d all be poisoned before we got the first shelter built,
and those same winds may have stripped away any breathable atmosphere,
meaning we’d have to live underground. It’s also possible that the planet’s orbit of
Proxima Centauri happens at such a cadence that one side of the planet
permanently faces the sun, meaning half of the planet is always daylit, and the
other is always in darkness.
But let’s stay hopeful. Let’s imagine that it’s a perfectly habitable place, with
warm winds and a liquid ocean and strange, vivid landscapes of rock and
vegetation and alien snow. Let’s go there!
First, the good news. Proxima Centauri b is only 4.2 light-years away. That
means that light, the fastest thing we know of, at roughly 186,000 miles per
second, would take only 4.2 years to streak from our planet to Proxima Centauri
b’s weird, wild shores. For photons, that’s a very short trip.
The bad news is that for humans, it’s a very long trip. We don’t travel at that
speed. Not even close. We’ll need much more time to get there. In fact, it’s so
much time that no one who has ever set foot on Earth will ever set foot on
Proxima Centauri b.
If we were to board a spacecraft and ride it from the outer edge of our
atmosphere all the way to Proxima Centauri b, you and I, who boarded the ship fit
and trim, chosen as we were from billions of applicants, would die before the
voyage reached even 1/100th of the intervening distance. It’s such an
outrageously long journey that a human life span is just a tiny fraction of the time
it will take.
Here’s the napkin math. At a speed of 20,000 miles per hour—the speed of our
top-performing modern rockets—4.2 light-years translates to more than 130,000
years of space travel.
One hundred thirty thousand years. This means that the time involved to reach
our closest exoplanet neighbor would crush us, and our children, and their
children into dust a thousand times over before anyone had a chance to breathe
alien air.
Could we put ourselves in some sort of coma for the journey, as the characters
in 2001, Alien, and Interstellar do? Trauma surgeons, experimenting with the
same concept that inspired Ted Williams’s family to freeze his head for possible
transplant in the future, are currently experimenting with procedures that can
revive a semi-frozen patient after two hours without a pulse. But we’re a long
way from freezing ourselves for as long as this might take. Using current
technology, anywhere from 900 to 1,300 human generations would pass on the
way to Proxima Centauri b before the ship arrives. Generations. So how will we
ever get there? A generation ship.
First proposed in varying forms by early rocket pioneers, science fiction
writers, and astrophysicists with a few beers in them, the general notion is this:
get enough human beings onto a ship, with adequate genetic diversity among us,
that we and our fellow passengers cohabitate as a village, reproducing and raising
families who go on to mourn you and me and raise new children of their own,
until, thousands of years after our ship leaves Earth’s gravity, the distant
descendants of the crew that left Earth finally break through the atmosphere of
our new home.
I once had dinner with an evolutionary biologist, and I asked him what it is
about Darwin’s theory that the human mind has most trouble seeing clearly. “It’s
the outrageously long periods of time involved,” he said without hesitation.
“We’re just not equipped to be able to imagine the amount of time evolution
takes.”
That inability to see evolution’s progress on the vast plain of time also means
that planning for ongoing communal life aboard a single spaceship is largely
beyond our natural gifts. I pat myself on the back when I get out ahead of
birthday presents for my children. The ability to plan an environment that will
keep my great-grandchildren alive and happy enough to reproduce with the great-
grandchildren of my colleagues is an entirely separate matter.
Think of the logistics: How will we create new clothes a thousand years after
the first boxes of cloth have collapsed and moldered away? Think of nutrition:
What crop lasts even 100 years, much less 100,000 years? And relationships: Will
assortative mating, the natural tendency to band together with people that most
resemble us, continue onboard, cutting into the genetic diversity of offspring and
wiping out that bloodline after a few succeeding generations? (Speaking of
factions, I have half a screenplay written about the tribal nightmare that such a
ship might become. Movie producers: I’m easy to find online. Hit me up. You
only need to build one set!)
Consider this: humans only spread across the planet roughly 140,000 years
ago. That’s when our species stood up, looked around, and accessed some higher
form of ambition and curiosity in deciding that it was worth risking everything to
leave what is now the African continent and look for game and water elsewhere.
Every human accomplishment you’ve ever read about—language, religion, the
Egyptian, Greek, and Roman empires—has happened since then. We’ve only
been “modern humans” for that long. And that means that a generation ship isn’t
just about finding and training the right one hundred people to seed a thousand-
generation journey. It’s about literally bottling the equivalent of our entire history
as a species for a repeat performance in captivity.
A generation ship is every sociological and psychological challenge of modern
life squashed into a microcosmic tube of survival and amplified—generation after
generation. Will the final generation be able to function well enough to even set
the ancient ship, rebuilt in flight over and over, down on the surface of humanity’s
new home? How do we possibly plan for this sort of time scale?
The idea of a generation ship felt like a pointless fantasy when I first
encountered it. But as I’ve spent the last few years speaking with technologists,
academics, and policy makers about the hidden dangers of building systems that
could reprogram our behavior now and for generations to come, I realized that the
generation ship is real. We’re on board it right now.
On this planet, our own generation ship, we were once passengers. But now,
without any training, we’re at the helm. We have built lives for ourselves on this
planet that extend far beyond our natural place in this world. And now we are on
the verge of reprogramming not only the planet, but one another, for efficiency
and profit. We are turning systems loose on the decks of the ship that will
fundamentally reshape the behavior of everyone on board, such that they will pass
those behaviors on to their progeny, and they might not even realize what they’ve
done. This pattern will repeat itself, and play out over generations in a behavioral
and technological cycle. That’s what this book is about.
I call this book The Loop, but there are, in fact, three loops I want to describe.
Think of a small loop nested inside a second, and the second loop nested inside a
third. Seeing each one clearly requires examining the others. And I’m going to
begin with the smallest loop, the one at the center of everything we do, one that’s
been spinning for tens of thousands of years, maybe even millions.
The innermost loop is human behavior as we inherited it, the natural
tendencies and autopilot functions that evolution gave us. That loop includes
everything from racial bias to our inability to accurately perceive risk. That’s the
first section of this book, because understanding the other loops requires
understanding how they rest on this one.
Once we understand that, we can look at the second loop. The second loop is
the way that modern forces—consumer technology, capitalism, marketing,
politics—have sampled the innermost loop of behavior and reflected those
patterns back at us, resulting in everything from our addiction to cigarettes and
gambling to systemic racism in real estate and machine learning.
The outermost loop is the one for which this book is named. It’s what I worry
is coming, a future in which our ancient and modern tendencies have been
studied, sampled, fed into automated pattern-recognition systems, and sold back
to us in servings we will be doubly conditioned to consume without a second
thought. It’s the loop that has already erased our conscious musical tastes and
replaced them with weekly suggestions from Spotify, based on what was served
last week. It’s the loop that samples our political tendencies and social appetites
and puts us together on Facebook with people whose patterns complement our
own, such that we become attached to a powerful and all-consuming group
identity, and have no patience for people outside of it. It’s the loop that looks
across all available job applicants and highlights only the ones that match the
characteristics of others who’ve done the job well, so that you, the recruiter, begin
to assume hiring can happen in a day or two, rather than weeks, without ever
leaving your home. The Loop is a downward tailspin of shrinking choices,
supercharged by capitalist efficiency, in which human agency is under threat from
irresistible systems packaged for our unconscious acceptance.
It’s happening already, but the real threat is what will happen over time. The
most powerful and lasting effects will take at least a generation to play out. Think
of it this way: we’re guided by unconscious tendencies, but we rarely detect it
when they are analyzed and played upon. Now throw pattern-recognition
technologies and decision-guidance strategies at us. And do it all in a society that
doesn’t have the long-term sensibilities in our policies and in our programming to
recognize and regulate something that will determine the future of the species.
That’s what this book is about.
I hope to describe here the challenge facing us as we build technology that will
alter coming generations of human behavior. Our behavior may be made more
efficient, more rewarding, more controlled in the current generation. We may reap
immediate economic benefits, and we may be able to do away with difficult and
sometimes terrible work in the short term. For every dangerous example I
describe here, there will be complex benefits connected to it. But can we look
with Darwin’s eyes ahead at the long-term possibilities of generation after
generation experiencing a collapsing cycle of choices, and changing for the worse
along the way? Can we somehow predict it, and act to correct it, even when the
rewards of doing so won’t be felt until you and I and our children are not just long
dead, but long forgotten? If we can’t, I worry that in a generation or two we’ll be
an entirely different species—distracted, obedient, helpless to resist the
technologies we use to make our choices for us, even when they’re the wrong
choices.
But I think we can fight The Loop. We just have to stand back, sketch out its
shape, recognize the ancient preconditions in all of us that make it possible, spot
the modern paths it takes into our lives, and then break its influence on our future.
Jacob Ward
Oakland, California
August 2021
Chapter 1
SINCE THE 1950S, when scientists began to revisit Pötzl’s ideas, researchers
have been trying to sort out just how much of reality we actually experience. And
they’ve found that the endless phenomena we observe through our senses—the
sweating beer and the job application on the table and the moon rotating across
the window and the distant sound of lovers quarreling and the warm breeze
through the curtains—are smashed and reordered and rewritten as our brain
makes the most efficient sense it can of each scene. Here are just a few examples
from a vast canon of these findings.
Let’s begin with vision. The study of sight has revealed that what we think
we’re witnessing—the unfiltered stream of objects and people and events we see
—is actually a wild variety of stuff, some of it perceived, some of it remembered,
all of it felt in a way we haven’t fully defined yet.
In 1992, a pair of cognitive neuroscientists, Mel Goodale and David Milner,
wrote a groundbreaking paper that sought to define this notion. It suggested that
we have two different ways of seeing an object in front of us: one way involves
consciously perceiving the object and making decisions about it, another
instantaneously guides the moves we make in relation to that object. They put it
more succinctly: “vision for perception and vision for action.” Writing in a British
psychology journal, they admitted that this “two-streams hypothesis” often made
them unpopular at parties: “The most difficult aspect of our ideas for many people
to accept has been the notion that what we are consciously seeing is not what is in
direct control of our visually guided actions,” they wrote. “The idea seems to fly
in the face of common sense.”5 Our actions feel voluntary, and therefore must be
under the direct control of our will, right?
Well, no. Goodale and Milner found, in carefully crafted lab experiments, that
our perception is not under direct control of our will. When they put an illusion in
front of their subjects, what the subjects reported seeing and what their hands did
were often dramatically disconnected. For instance, remember that fun-house
illusion in which a face is carved into the wall but looks as if it’s protruding out
from it? The one where you walk past and the face seems to follow you? Well,
when they showed people that illusion, called a “hollow face,” the conscious
mind was usually fooled, and thought a sculpture was turning its head to track
them as they crossed in front of it. (I remember this very thing scaring the bejesus
out of me at a haunted house as a child.) But when they asked the same person to
flick a small bug-like target off that face, their hands reached inside the hollow
space that had fooled the eye, past the boundary of the face the conscious mind
had seen, and flicked the bug at exactly the right spot. As the researchers describe
it, these were “two parallel visual systems—each constructing its own version of
reality.”
Other senses also seem to be unconsciously received and assembled in the
mind as well. The human talent known as stereognosis—the ability to identify an
object by touch alone—helps give us incredible manual dexterity. But the
processing of texture, weight, and the rest of it seems to take place in the parietal
lobe of the brain, a long way from the portion that controls our hands. That may
be why stereognosis begins to crumble in patients suffering from Alzheimer’s—a
disease that affects the parietal lobe—whereas patients with other forms of
dementia retain that ability. Your fingertip doesn’t experience reality. The brain
assembles the things you’ve touched into its own imaginary rendition of them.
And hearing isn’t just a means of identifying a sound and triangulating its
position—it also seems to be bound up in other mental tasks, like your brain’s
ability to remember things in the proper order. Researchers call the idea “auditory
scaffolding”: that hearing supports the acquisition and retention of things like
long lists of words or numbers. Studies of children who are deaf have found that
although all their other senses are intact and developmentally they’re otherwise
on track, they tend to do worse than hearing children do on certain tests of
sequential abilities, even ones as simple as tapping their fingers together in a
prescribed order. We aren’t consciously hearing the raw sounds of the world. The
brain is using certain aspects of our hearing to put together its own soundtrack.
So if perception isn’t a matter of simply seeing, touching, or hearing reality,
what is doing the postproduction work right before it’s all shown to us? Maybe
memory is involved. The neuroscientist David Eagleman spent a decade
experimenting with the mental perception of time. One of Eagleman’s
experiments would have made even Pötzl spit out his tea.
Eagleman and his graduate student Chess Stetson were looking for a way to
test the common report that time “slows down” during brief, dangerous events. In
my twenties I was headed downhill on my bike when the car next to me suddenly
nosed across the road into a parking space. My bike stopped dead against the
car’s turned front tire, and I flew into the air across the hood. I distinctly
remember having an entire conversation with myself as I traveled through space.
“You don’t have health insurance! You didn’t mail in that COBRA application
after leaving that job!”
“Yes, I did, actually. I remember stamping it and mailing it last week.”
“Oh yes, you’re correct, and that’s great news, because it looks like you’re
about to be pretty badly injured.” And then I slammed into the asphalt and slid a
few feet, breaking my wrist. I remember sitting up, dazed and bloody, looking at
my bent bike. It all felt as if my brain had affected the flow of time for my
benefit.
Eagleman and Stetson worked to find out whether the brain does, in fact, slow
things down. Neuroscientists had already determined that when the average brain
receives two images within one hundred milliseconds of each other, it fuses them
together, and we see one image. So Eagleman and Stetson built an enormous
wristwatch that displayed random numbers just fast enough—less than one
hundred milliseconds apart—that subjects couldn’t consciously see them. First,
the subjects were shown the numbers in a laboratory setting. They reported that
the numbers were just a blur. Then Eagleman and Stetson dropped these people
off a crane.
No, really, they marched their subjects to the top of a carnival ride, equipped
with a helmet and the big wristwatch, dangled them from a cable, and dropped
them fifteen stories into a net.
Here is how Eagleman summed up his findings in a 2009 essay:6
The result? Participants weren’t able to read the numbers in free fall any
better than in the laboratory. This was not because they closed their eyes or
didn’t pay attention (we monitored for that) but because they could not,
after all, see time in slow motion (or in “bullet time,” like Neo in The
Matrix). Nonetheless, their perception of the elapsed duration itself was
greatly affected. We asked them to retrospectively reproduce the duration
of their fall using a stopwatch. (“Re-create your free-fall in your mind.
Press the stopwatch when you are released, then press it again when you
feel yourself hit the net.”) Here, consistent with the anecdotal reports, their
duration estimates of their own fall were a third greater, on average, than
their recreations of the fall of others.
How do we make sense of the fact that participants in free fall reported
a duration expansion yet gained no increased discrimination capacities in
the time domain during the fall? The answer is that time and memory are
tightly linked.
ILLUSIONS
TWO SYSTEMS
IN 2013, I stood under the burning sun in Homestead, Florida, watching robots
fall over.
They were attempting to navigate an obstacle course on a sun-bleached motor
speedway where I and dozens of other reporters from around the world squinted
in the heat, even through our sunglasses. The assembled machines were intricate
contraptions. One of them could scurry around on all fours across a field of
rubble, then rise onto its hind legs to grasp at a door handle. (That one managed
to play on both my fear of insects and my fear of home invasion.) Another had a
sort of stooped, careful stance that reminded me of my grandfather shooting pool.
A third had a sleek white and gold exoskeleton that could someday qualify it to
check IDs outside a nightclub. They all looked very impressive. The products of
the best minds in civilian, military, and academic robotics from all over the world,
these creations were part of a tournament called the DARPA Robotics Challenge.
I’d flown in with a camera crew to cover the event, promising my bosses at Al
Jazeera that this would be made-for-TV mechanical drama at its best. But that
morning I realized I was going to have to apologize, because an early task was
crushing every robot.
They were all trying to get up a ladder.
DARPA is the Defense Advanced Research Projects Agency, a think tank for
the Defense Department. It exists to pull far-out ambitions into reality. An earlier
iteration of the agency pursued a networked information system that could
survive a nuclear attack and wound up inventing the Internet. These days, in
addition to offering hundreds of millions of dollars in grant money, it also holds
highly lucrative competitions. This particular competition in Florida was an
annual competition, a yearly federal goosing of the robotics industry into
producing an all-in-one robot helper for disaster work. (Although of course a
robot capable of these tasks could have other, more military purposes as well.)
This same competition format previously produced a functional self-driving car,
after several years during which the entrants careened off roadways and crashed
into barriers. But this year it didn’t reach such heights with the robotic assistants
that DARPA imagined someday putting out fires, handling nuclear material, or
carrying lunch across a minefield.
The trouble was that these robots were faced with human tasks, in human
environments. To an outside observer like me, who knew next to nothing about
building a robot, the jobs looked easy. Open a door. Use a drill. Get out of a car.
The ladder, if anything, should have been the simplest. To the human eye, a ladder
consists of a perfectly reasonable number of variables. The handholds are smooth
and inviting, the rungs are consistently spaced and wide enough to comfortably
hold a foot, and when balanced right, it’s very sturdy. This particular one, built to
hold thousands of pounds of robot, was more like the stairs one might ascend to
give a speech—a kid could easily get up it—and yet it absolutely murdered these
robots.
The problem is data. Even though the ladder looks simple, the incoming
variables are in fact overwhelming. A robot approaching the staircase is presented
with reams and reams of unfamiliar information. And it has to consider each item
individually. Where is the first rung? Is it visually separated from the ground?
How about the second rung? Can one infer from the relation between rungs one
and two what the angle of the whole thing is going to be? Does the railing follow
the same angle? And where does it end? The tiniest interaction with the ladder
turns out to be incredibly complicated. And no one task can be done instinctively
by a robot, as you and I do so many things. The robot cannot unconsciously throw
a foot the same distance each time, feeling for the foothold with a toe, and the
actuators that keep it in balance don’t coordinate automatically with the signals
coming from its lower extremities, as mine do. The robots have to consider each
and every factor, no matter how small. It all has to be processed consciously, if
that’s the word. And they just didn’t have enough conscious bandwidth to handle
the situation. Time and again, the robots in Florida went over—some backward,
some sideways, some headfirst into the rungs—and a helmeted member of the
robotics team in question had to save them at the last moment, desperately
hoisting the falling robot off the stairs on a belay rope, everyone on the team
cursing into the sky or at their own feet as their mechanical champion dangled in
the air.
Even though I’m arguing that we should be worried about the advanced state
of our technology, and these robots were pathetically bad at ladder-climbing, I’m
telling you about their laughable failings for two reasons.
First, they represent the grand ambitions of technologists to build sophisticated
stand-ins for the human mind and body. In two decades covering technology as a
reporter, I have over and over again been shown technology whose creators talk
about replacing or supplementing humans in all manner of tasks. But the robots of
the DARPA Challenge, and their painful efforts to stand in for us, are the gap
between what we want technology to do, and what it can actually accomplish.
These robots are fascinating and frightening, but they are not at all ready to
replace humans, no matter how desperately we might want to hand one a fire hose
and tell it to start saving lives. It may be that someday they will fix broken screen
doors and extinguish fires for us. But our desire to give them those jobs—and our
unconscious willingness to believe that they can do them—doesn’t match the
robots themselves. I went to Florida expecting miraculous artificial beings. I left
Florida with my expectations reset. I’ve tried to carry those expectations with me
ever since.
The other reason to consider these robots here is that their failure to do what
we do tells us something fundamental about what we do. Just as Kahneman and
Tversky showed that we process uncertainty in an unconscious way, it turns out
we process the rungs of the ladder, a soup spoon, and our shoelaces outside our
consciousness. The ability to see and manipulate these objects takes place outside
our waking minds. It’s a system that has kept us alive for millions of years
because it takes the load off our consciousness, freeing it up to do more important
things, like invent new ladders and spoons and shoe closures. It’s an extraordinary
gift. And it’s important to understand it, to understand that we haven’t yet built a
stand-in for it, and to understand that we are nonetheless geared to unconsciously
offload difficult cognitive tasks to automatic systems, whether they be our
emotions or a collection of shiny but unreliable robots, even when they’re not up
to the job.
CLUSTERS
WHAT ABOUT THE beauty of our natural state? Shouldn’t we be striving to live
as we did in ancient times, gathered peacefully around a fire, sharing our food?
Considering that so much of our modern world seems to be disorienting the
ancient systems we use to navigate it, shouldn’t we be steering ourselves back
toward the pure, honest, sustainable ways of ancient life?
“No no,” says Mahzarin Banaji. “Our natural state is terrible, horrible. You’d
never want to go back to that.”
Banaji, a professor of social ethics in the Harvard psychology department, is
an electric personality, crackling with humor behind thick, colorful glasses. But
her work, studying the ancient biases we don’t recognize in ourselves, is a
bummer. She finds her own path to optimism, though, and often begins her
speeches to large audiences with a spectacular bit of dark humor. “I’d like to
congratulate you,” she’ll say, “on the sheer diversity of this crowd. You all come
from wildly different backgrounds, you are drawn from different genders, ages,
social status, races, religions. In the tens of thousands of years in human history,
you are the first group this large and this mixed up in which no one is going to die
before I’m finished speaking. There won’t be a riot, no one will be burned for
being different. It’s amazing. Congratulations!”
In 1996, as a graduate student in psychology at the Ohio State University,
Banaji was helping to design experiments about bias. At the time she thought
very highly of herself as a fair-minded person. And if anyone’s life story suggests
modern conscientiousness, an immunity to ancient prejudices, it’s Banaji’s. She’d
grown up Zoroastrian in a Parsi community in India, mentoring other students
from the age of five in the small school her mother and aunt ran in their home
before rocketing through a series of schools and universities and winning a
fellowship to study in the United States. “I figured I was about as unbiased as one
could be, and I thought very little of certain parts of society and their biases,” she
says.
And so she sat down one day to take an early version of the bias test she and
her colleagues had designed, feeling perfectly ready to have her self-image
reflected back at her. Up to that point, psychologists had largely assumed that
they could measure bias by asking people to volunteer their own attitudes toward
other people. But that sort of thing only revealed what people thought their biases
might be. The test in front of Banaji was different. She was part of a group trying
to design something that would draw out “implicit bias”: unconscious attitudes,
ones of which we’re not only unaware, but we also might be horrified to discover.
It’s a common term today—presidential candidates have used it on the campaign
trail, and it’s the subject of countless corporate training programs—but at the time
it was a revolutionary idea. The test asked Banaji to look at a series of faces, and
as each appeared, press one of two keys to identify the face as Black or white. At
the same time, the test asked her to use the same two keys to label occasional
words as either positive or negative, one key assigned to each. The test had seven
stages, during which it kept switching up which keys did what; one key would be
used for both white faces and positive words, then in the next test it would be
used for white faces and negative words. The other key would be used for Black
faces and positive words, then Black faces and negative words. It’s laborious to
describe, and it’s a dizzying test, just tedious enough to cause the subject to lose
focus, and something about that, along with using the same finger to denote both
a racial identification and an emotion, teases out long-held, instinctive
connections we make between the two. In the end, the test told Banaji she had “a
strong automatic preference for European Americans over African Americans.”
“I had two reactions,” says Banaji, now a towering figure in the study of bias.
“The first was to assume that the test must be wrong. I mean, if my two fingers on
a keyboard can’t produce the result I’d expect about my own attitudes, the test
must be flawed! And then, when I discovered that the test wasn’t wrong, I was
embarrassed! Mortified!”
Banaji and her colleagues Tony Greenwald and Brian Nosek put the test online
in 1998, hoping that perhaps five hundred people might take it in the first year.
Instead, forty-five thousand people took the test in the first month. That
popularity raised the question of selection bias. They worried that only liberal
people who consciously wanted to fight prejudice would take it, skewing the
results. But that quickly faded as people of all backgrounds poured in, day after
day. “I remember one day looking at it, and we noticed 400 people from Topeka,
Kansas, had taken the test,” Banaji says. “And we realized, ‘Oh, it was a school
that sent a large number of kids to us.’” Accounting firms, military units, museum
staffs, all came through. “Neo-Nazi groups used to visit in the old days—we used
to get messages from them.” And the flow of data has not slowed. “For these 22
years, thousands of people come to it every day,” Banaji says.
Taken by more than 30 million people since it first went live, the Implicit
Association Test, or IAT, uses simple keyboard mechanics to measure our
instinctive attitudes about gender, age, race, sexual orientation, disability, body
weight. And because the test has more or less remained the same for so long, as
new participants take it, the IAT at this point represents a comprehensive
longitudinal survey of how attitudes have changed over the more than two
decades since it began.
The good news is that certain attributes—like sexual orientation—are losing
their stigma. “Perhaps it’s that people often know someone who is gay, they can
relate to them personally, they’re a family member, a child,” Banaji says. Senator
Rob Portman of Ohio had been, for most of his career, staunchly opposed to
legalizing same-sex marriage. But then his son Will came out to him and his wife.
“And that launched an interesting process for me, of rethinking my position on
the issue,” he told CNN correspondent Dana Bash in a 2013 interview. “I now
believe that people have the right to get married.”
Banaji waves off Portman’s change of heart. “I wouldn’t extend any kudos to
Senator Portman,” she scoffs. “Apparently you need a personal experience to
change!” That’s not going to make any difference in other areas of bias, she says.
“Your son is never going to come back after his first semester at Yale and say to
you ‘Dad, I’m Black.’”
And unfortunately, Banaji’s data shows that racial bias, while it has fallen ever
so slightly in the last decade, is still more or less as it was. Banaji cuts a straight
horizontal line through the air with her hand. “Race is like this,” she says. “I think
people would be very surprised to discover that our racial biases really haven’t
changed at all.”
But what about seminal cultural and historical moments? The Cosby Show?
Oprah? The election of nonwhite candidates like Kamala Harris to the highest
offices? Sadly, they don’t move the needle. “The day after Obama was elected we
were looking at the test results to see if anything moved,” Banaji says. “And no.”
This isn’t to say that our conscious attitudes can’t be changed by a singular event.
“Often people will report an individual experience with somebody of another
group that changed their minds completely. This is why we encourage foreign
travel. But I think for what we are talking about,”—long-term, large-scale,
society-wide changes of attitude—“this bean-counting little machine”—your
brain—“is saying ‘x goes with y.’”
As we’ve seen, our brains use long-term patterns to take cognitive shortcuts.
And Banaji explains that changing those patterns requires huge new amounts of
counterfactual data. “What we’re talking about here is something that operates at
a level we’re not conscious of,” she says. “‘What’s the probability that when I
stop at this newspaper stand in New York City that the owner will be South
Asian?’ And because the probability is so high, we will make that association. For
it to go away, we would have to show them many, many, many instances of
walking up to a newspaper stand, and having a quite different person be there.”
Awful as it seems today, there is a reason for all this. She describes bias as a
lifesaving adaptation to a brutal world, one in which we lived in small clusters,
where any outsider might be a threat. Being able to instinctively and rapidly
identify someone as an outsider was a wonderful thing. And being able to react
appropriately based on assumptions about one’s surroundings through repeat
exposure to the color of ripe fruit or the movement of a predator in the firelight
was a time-saving way of staying alive and protecting our communities.
Researchers even believe there may be psychological mechanisms at work that
help us avoid disease by avoiding outsiders, sticking to our own routines, and
being generally socially conservative, a set of unconscious strategies known as
the behavioral immune system.1
But today, our relentless statistical bean-counter minds are a terrible
disadvantage to us when it comes to living according to our new, higher ambitions
as human beings. Our lives have changed; our unconscious habits haven’t. We
aren’t built to treat new people fairly, to inquire about their lives and invite them
into ours. Millions of years of evolutionary pressure served to crush open-
mindedness, because open-mindedness would have gotten us killed. Any shred of
it we have is effectively made up. We put equitability on like a costume, a smooth
modern aspiration draped over sharp, dangerous, ancient programming.
I once interviewed Lord John Alderdice, who grew up during the troubles in
Belfast, and was part of the political movement that built the Good Friday
Accord, the result of years of negotiations for peace in Northern Ireland. He’s a
member of the House of Lords and now studies conflict around the world. He’s
passionate about the work that wildly different people do to find common ground,
but as a practicing psychiatrist and someone who has spent decades trying to
facilitate that work, he believes it runs contrary to our natural instincts.
“The whole process of globalization, the freedom to trade and travel and move
about, not just you going to other countries, but them coming to you—there’s a
certain percentage of the population, perhaps 15% or so, who like that idea,” he
says. “And these in the main are people who inhabit newspapers and academic
institutions. They feel happy and comfortable about it all. But what they don’t get
is that most people don’t feel that way. They just want to holiday and come home.
What they don’t want is other people coming, not for a holiday, but bringing with
them their cultures, and doing so at a rate of change they can’t cope with.”
Consider just how new it is to be able to think the way we do. Our closest
common ancestor with modern-day primates walked the earth perhaps 30 million
years ago. We began to resemble our current form roughly two million years ago.
But it was only perhaps 200,000 to 70,000 years ago that meat-eating and heaven
knows what else pushed us to wonder what lies beyond our bodies and our
immediate needs. The neocortex, as neuroscientists call it, is the part of the brain
that gives us the ability, acquired very recently in the evolutionary sense, to think
beyond ourselves—to not just pick up cues from others of our kind, but put
ourselves in their shoes, stay up late worrying about our future and past as
individuals and as a group, empathizing and negotiating and coordinating with
each other on a constant and largely peaceful basis.
Our ancient selves developed the ability to quickly recognize, ostracize, and
even punish difference. Those ancient humans could not have done what you do
each day. And we’re not just talking about the big stuff, the peace negotiations
between warring peoples that experts like Lord Alderdice help to facilitate. We’re
talking about any of the trappings of your daily life. Ancient peoples could not
have woken in a home respected as their private sanctuary and sat down to work
all day with a wildly varied group of other humans. Someone would murder
someone, as Banaji likes to point out. They certainly couldn’t have voted
peacefully for leaders they’d never met and will never meet, sued one another in
court, agreed to treat certain pieces of paper as having value, or placed the smaller
of their two carry-ons under the seat in front of them. The trappings of our lives
are all astoundingly new and alien, and when we see the mismatch between our
cognitive equipment and the work we do with it, and then realize the degree to
which our ancient biases predetermine how we negotiate the corridors of modern
life the way TN negotiated the hallways of Beatrice de Gelder’s lab, it becomes
clear that this balancing act is very fragile indeed.
Kahneman wrote in 2003 that his work on System 1 and System 2 drew from
two broad hypotheses about human decision-making. “The first is that most
behavior is intuitive, skilled, unproblematic, and successful.” Automatic
decisions, after all, have kept the species alive all this time. But the second
hypothesis is where we get into trouble. “Behavior is likely to be anchored in
intuitive impressions and intentions even when it is not completely dominated by
them.” You and I are more or less always influenced by these unconscious
systems even when we’ve exercised some rationality.
Still, haven’t we created ethical standards and governing principles and
legitimized organizations that correct for those outdated instincts? Can’t we
reason our way out of this? I believe it’s one of the miracles of modern life that
we can agree as a society to stay in our lanes on a highway, not to mention that
we can go to court peacefully when we’ve been wronged. But sadly those systems
don’t mean we’ve left our ancient instincts behind. Just as Alexander Todorov and
his colleagues found that we vote with those instincts, even when we think we’re
voting with our higher functions, the Yale psychologist Yarrow Dunham has
found that in-group/out-group biases are so ready to pop to the surface that he can
draw them out of children as soon as he meets them.
Dunham is a friendly, open-hearted sort of guy, quick to laugh, warm with
strangers. Not exactly the kind of person you’d expect to have spent years
uncovering powerful discriminatory tendencies in children. His lab at Yale
directly tests the ways human behavior is, as Kahneman wrote, “anchored in
intuitive impressions.” I watched an experiment in which he assigned children to
arbitrary groups, and then had them share what they assumed about their own
groups and others. It’s terrifying to witness.
In his experiments, Dunham asks small children, between three and six, to
spin an arrow on a wheel. The arrow lands on either green or orange, and
Dunham presents the child with a T-shirt in that color, pronouncing them a
member of the green or orange team. By that point, he tells me, “I can have them
making outrageous assumptions about each other within 90 seconds.”
I watched him do this several times. He’d begin by testing the children’s
assumptions about their own team and about the other. Which of these two kids,
green or orange, is most likely to give you candy? Which of these two would you
want to be friends with? The results were immediate, and consistent, and awful.
He showed one girl, dressed proudly in her new orange T-shirt, an image of an
ambiguous scene. In it, a child in orange lies sprawled beneath a swing set while
another child in green stands behind the swing, looking concerned. “He pushed
her off the swing,” the girl told him. But later, when a boy in a green shirt
examined the same scene, he told Dunham that the girl in orange must have
fallen, and that the boy in green was about to help her up. Same scene, same facts,
interpreted opposite ways.
“They start out with just a small bias in favor of one group over the other,”
Dunham explains. But over time, as he shows them more of these scenes, their
biased interpretation of the scenes begins to become a body of evidence in their
minds. And that’s when assumptions really start to lock in. “When I later ask
them about the group, they have a little stockpile of what they think are facts to
draw on. ‘Oh, well, members of that group are just mean, they’re pushing people
over all the time, they’re stealing people’s money.’ So now they have essentially
evidence in favor of their bias, and now you can imagine this leading them even
more to think that that initial bias is justified.”
Now imagine playing those assumptions out over a life span, Dunham says.
“Over time the accumulated weight of all of that is a very entrenched view of
your own group as positive.”
Dunham’s work is part of a decades-long effort to find out why we so quickly
give ourselves over to biases about even the most arbitrary and petty of
groupings. Turns out we’re just built for it, somehow. In the 1950s, the social
psychologist Henri Tajfel, a Polish Jew who survived the war concealing his
background in a series of German POW camps, sought to find out why even
ordinary Germans had gone along with Nazism and its horrors. By 1970, he was
conducting experiments at the University of Bristol in England on the notion of a
“minimal group paradigm.” The idea was that it took nothing more than
membership in a group—people connected by anything, from similar Christmas
sweaters to nicknames—for the members to begin favoring one another over
outsiders. Ever waved at a passing driver because they happen to be in the same
kind of car you are? That’s the idea here.
It turns out Tajfel wasn’t immune to the effects he studied. A pair of
researchers put together an investigation in 20192 that found Tajfel was not only a
serial sexual harasser of the women in his lab, but remained “resolutely
uninterested” in applying his findings about in-group/out-group preferences to
gender, and neither treated women as intellectual equals nor pulled them into
meaningful collaborations. In Europe, the organizers of a prominent academic
award named for him recently announced they’d be renaming the prize.
But his findings about a minimal group paradigm are as strong today as his
unfortunate personal demonstration of it. Hundreds of experiments have
reproduced the same dynamic. Slap us into random groups and we’ll assume ours
is superior to another. (As the psychologists Jeffrey Lees and Mina Cikara have
found, perceptions of what the “other side” believes are also often inaccurately
pessimistic, even on subjects where both sides largely agree on the issues.) And
as Dunham explains, that tendency camouflages itself in our minds as objective
judgment.
GUIDANCE SYSTEMS
WHAT ARE THE specific human instincts being used to manipulate us? What’s
the way in? As early as the 1970s, Kahneman and Tversky were working on a
new concept for measuring human instincts around risk and uncertainty.
In their 1979 paper “Prospect Theory: An Analysis of Decision Under Risk,”
Kahneman and Tversky identified a tendency that upended decades of thinking
about risk analysis. Rather than making rational decisions based on a thorough
consideration of the straight probability of a good or bad thing taking place, they
theorized, humans instead weigh gains and losses on a strange curve: our hatred
of loss makes us wary of even statistically smart gambles, and we’ll do almost
anything to convert a tiny bit of uncertainty into a sure thing. This was the
beginning of what Slovic would go on to show was an outsourcing of decisions to
our emotions.
Kahneman and Tversky deployed a devilish sort of question to tease out this
tendency.
A FEW MONTHS after Patrick described his heroin overdose to me, I was
invited to join a loose dinner party elsewhere in San Francisco thrown by a group
that called itself BTech, short for Behavioral Tech. It was a monthly gathering of
young neuroscientists, behavioral economists, and other specialists in human
decision-making who now work in the tech industry. The gathering of perhaps
fifteen or so that night included someone from Lyft; someone from Fitbit; the
creator of a financial-advice chatbot; and Nir Eyal, a Stanford MBA who had sold
two companies, written a best-selling book called Hooked: How to Build Habit-
Forming Products, and was now a consultant successful enough to have built out
the white-walled former storefront in which we were all seated. I wrote to the
group ahead of time to inform them that I’d be there, that I was writing a book,
and that whatever they said might wind up in it. Over Coronas and Indian food,
we listened to a pair of newly graduated neuroscience PhDs, T. Dalton Combs
and Ramsay Brown, talk about their company, called Dopamine Labs, through
which they were selling their academic knowledge of dopamine release to help
make apps more addictive. They were there that night to talk about the market
potential, and to offer their services.
Combs and Brown began the evening showing off an app they’d built, called
Space, which could interrupt the feedback loop one gets from opening social
media apps by enforcing a few seconds of mindfulness before the app would
open. But by the end of the evening, it was clear they were happy to sell these
same principles in the service of almost any company. As Combs told us, what’s
so great about the human mind is that if you can just manipulate someone’s
habits, their consciousness will invent a whole story that explains the change of
behavior you tricked them into, as if it had been their idea all along. It makes
what they were calling “behavior design” so easy, he told us.
Now, the room, I should point out, was full of good intentions. Many of the
apps had to do with fitness and saving money and healthy meal plans. But there
was an open way of discussing neuroscience and behavior change and habit
formation that felt like a loaded gun. And when one member of the group asked
the Dopamine Labs founders whether there was any company they wouldn’t be
willing to work for, Combs said, “We don’t think we should be the thought police
of the Internet.”
Dopamine Labs’ presentation centered on the idea that companies can use our
hardwired habit system—the same stuff studied by the likes of Kahneman,
Tversky, and Slovic, what the two neuroscientists described as a chain of triggers
and actions and feedback and rewards that can be manipulated the way one might
pull levers to drive a crane—to get people to turn over information about
themselves through apps. These were folks trying to understand how to
manipulate—or, hell, create—the habits of their customers. And eventually we
wound up talking about drug addiction.
At one point in his presentation, Combs said that humans form habits through
contextual triggers. “If you take a habituated drug user who has recovered, let’s
say it’s someone who used to do cocaine in nightclubs—” He checked that we
were with him. Forks and beer bottles were frozen mid-journey. “—Let’s say you
bring them to a nightclub, and you show them baking soda, and you even tell
them it’s baking soda, they’ll still want to snort a line of baking soda.” (That’s me
outside the taqueria, I know now. That’s me inside any dark, wood-paneled bar.)
Someone in the room asked whether everyone is equally vulnerable to drug
addiction. I felt relieved that it had come up. After all, we’d been talking for
nearly three hours about the universal mechanisms of the human mind and how to
take advantage of them. At this point there seemed to be a collective realization
that we were playing with live ammunition. We’d clearly begun to push against
the ethical boundaries of all this behavior design. But somehow the people in the
room at that point managed to elevate themselves above the everyday people they
imagined would use their apps, drawing distinctions that conveniently left those
of us gathered in the room above the water line. And Eyal jumped in with a
sweeping statement. “Let’s be honest: the people in this room aren’t going to
become addicted. If you were injected with heroin a hundred times”—he gestured
at Combs, and then at the rest of the room—“you’re not going to become addicted
to heroin.” He seemed utterly convinced of this. “Unless you’ve got some sort of
deep-seated pain in you,” he says, “it’s just not going to happen.”
I didn’t share my opinion that pretty much any mammal injected with heroin
one hundred times would likely become addicted, but I did mention to him that
opioid overdose kills more people than car accidents these days, that at that time
it was the number-one cause of death by accidental injury in the United States,
that people from all walks of life were falling into opioid addiction at a rate never
seen before in American history. But he was unshakeable. “The people who
become addicted have pain in their lives. They return from war, or have some
other trauma, and that’s what gets them hooked,” he told me calmly. “Only two to
five percent of people are going to become addicts.” Forks and bottles began
moving again.
Combs and Brown went on to appear on 60 Minutes a few months after I met
them, and shortly after that they renamed their company Boundless Mind. Soon
Arianna Huffington’s Thrive Global bought them, and Combs was their head of
behavioral science until October 2020. He now runs a metabolic-fitness company.
Dopamine Labs does not appear on his LinkedIn profile.
Eyal went on to write a book called Indistractable: How to Control Your
Attention and Choose Your Life. It seems like a dramatic reversal for a guy who
wrote a marketing handbook called Hooked, but it’s not. His thesis in
Indistractable is that it’s up to individuals to learn self-discipline, that technology
is not addictive, that self-control, not regulation, is our pathway forward. This is
someone who used his understanding of psychology to establish and sell large-
scale tactics of persuasion and habit formation. And yet he had somehow
convinced himself—and, seemingly, most of the people there that night—that the
people making this tech are somehow not only able to resist what we had just
spent the evening establishing is an eons-old tendency of the brain, but that it was
all right for that rarefied group to deploy products that played on those
unconscious tendencies in the rest of us. Eyal went on to tell Ezra Klein in a 2019
podcast appearance, “The world is bifurcating into two types of people: people
who allow their attention to be manipulated and controlled by others, and those
who stand up and say, ‘No, I am indistractable. I live with intent.’”
When I watched the presentation from Dopamine Labs and heard Eyal’s
theories that night, I hadn’t yet met Robert Cialdini or Paul Slovic. I hadn’t yet
met Wendy Wood, and she was still a year away from publishing a 2018 review
of habit-formation science with the business professor Lucas Carden, in which
they looked across the field and concluded that people are not in control of their
choices. Wood and Carden compared our vulnerability to forming unconscious
habitual associations (between the smell of a grill and ordering a burger; between
the feeling of taking our work shoes off at the door and the first pull of an after-
work beer) to the inexorable pull of a rip current. The trick, they wrote, is to not
get in the water. “We now know that self-control involves a wide range of
responses beyond willpower. To be successful, people high in self-control appear
to play offense, not defense, by anticipating and avoiding self-control struggles.”5
And when I sat there with those young entrepreneurs, earnestly talking
through the ethics of addiction and design, I hadn’t yet met Kathleen Wilkinson,
who lost herself in an innocent diversion, costing her and her husband their last
years together. I hadn’t seen the ways that our most ancient instincts were being
built into emotion-shaping and decision-guidance systems for vast profit. I hadn’t
yet discovered that our unconscious tendencies are being amplified and shaped
and mutated for profit, by accident and sometimes even on purpose. But even
back then, I was frightened by the conversation I witnessed that evening, and,
knowing what I know now, I still am.
Chapter 6
WE’VE SEEN HOW the first loop works, the unconscious decision-making that
drives so much human behavior, and we’ve begun to look at the second loop, the
ways that technology-driven businesses are working to sample our unconscious
decisions and toss them back at us. As part of that, I want to step back and look at
how we not only came to pursue these sorts of manipulative technologies, but
why we so readily believe in them and trust the guidance they offer.
The truth is that in our history we’ve come under the direct supervision of
guidance systems before. And in the earliest iterations of them, they came to us in
the guise of lifesaving technology.
Between the end of World War II and the 1980s, organizations like the RAND
Corporation, the Cowles Commission for Research and Economics, and Stanford
University’s Center for Advanced Study in the Behavioral Sciences (where I
spent a year as a fellow oblivious to this part of its history) attracted bright young
thinkers from universities across the country who spent their time on the most
pressing mission of the era: how to avoid nuclear war. In their brilliant book How
Reason Almost Lost Its Mind, authors Paul Erickson, Judy L. Klein, Lorraine
Daston, Rebecca Lemov, Thomas Sturm, and Michael D. Gordin describe a long,
strange period during which these thinkers behind “Cold War rationalism” held
military leadership, academia, and political theorists in their thrall. As we face the
prospect of a world run by pattern-recognition technologies, we should closely
study this period.
It began with a triumph of logistics over death itself. In the spring of 1948,
American, French, and British forces, each occupying an agreed-upon zone of
defeated Berlin, were locked in a spiral of growing tension with the Soviet forces
there, who watched their Western counterparts with increasing alarm. Under the
terms of the Yalta and Potsdam conferences, Berlin fell deep inside Soviet
territory, yet the city was divided four ways. The Americans and British
announced plans to merge their zones, with the French soon to follow. Then the
Allies announced a new currency for their portion of the city, which would
devalue the Soviet-backed currency, and that’s when the Soviets imposed a
blockade of air, rail, and road traffic to the Allied zones, cut off their electricity,
and waited for starvation to drive the Allies out of Berlin.
But on June 26, 1948, RAF and USAF planes under the joint command of the
Americans began to roar in and out of the Gatow and Templehof airports in West
Berlin, packed with food and supplies. These were the first of nearly 300,000
flights that filled the sky above Berlin for more than a year. And because the
pilots followed disciplined procedures, remaining at designated altitudes,
navigating precisely mapped corridors back and forth, and taxiing with rigorous
precision, they managed to form a lifesaving, uninterrupted circulatory system for
the entire city. The flights and the unloading and loading were so efficient that at
one point aircraft were landing in Berlin every sixty-two seconds. The Soviets
had assumed that without trains and trucks the Allied half of Berlin would starve.
Instead, within eleven months the USSR had agreed to lift the blockade, and the
foreign ministers of the four nations were sitting down together.
The Berlin Airlift, a revolutionary rules-based system that had saved an entire
city and fought off the influence of the Soviet Union, set Cold War rationalism in
motion. This new institutional enthusiasm sprang from a movement inside the US
military, one accelerated by World War II, to apply computing resources and data
analysis toward creating new efficiencies and strategies. Mathematicians at the
Pentagon had already formed the Project for the Scientific Computation of
Optimum Programs, or Project SCOOP, tasked with using the punch-card
technology of the time to build new, data-driven systems. As Erickson, Klein, and
their coauthors describe it, the Project SCOOP team managed “the ingenious
design of an algorithm for deriving economically rational decisions from an
equation system that included an objective function maximizing, for example, the
tonnage delivered to Berlin.” But because they didn’t yet have the computing
power necessary to execute such an algorithm, “the only optimal program that
Project SCOOP could determine was for the least cost diet that could provide the
essential nutritional needs of an active, urban economist weighing seventy
kilograms.”1 The system had to work toward the simplest possible goal, which
was the calorie needs of the average man inside Project SCOOP—a primitive but
satisfactory objective function.
Inspired in part by Project SCOOP and its success in Berlin, a new field of
rationalism exploded. Researchers like the economist Herbert Simon, the
mathematician Anatol Rapoport, and the economist-turned-arms-control-expert
Thomas Schelling began their careers studying organizational psychology. But
geopolitical forces and military funding pulled the field toward questions of how
entities like the United States and the Soviet Union might avoid escalating
tensions to the point of mutual annihilation. What those researchers were hoping
to accomplish, it seems, was a sort of deprogramming—or maybe even
preprogramming—of international relations, using formulas adapted from
experiments like the Prisoner’s Dilemma, in which two participants must decide,
in isolation, whether to cooperate with or betray one another. The Cold War
rationalists appeared to believe that with disciplined analysis and creative
experimentation, it might be possible to boil down into a predictable set of
patterns all the complications that gave rise to international incidents like the
Cuban missile crisis. Understand those predictive patterns, and one could write
the choreography necessary to avoid them. Erickson, Klein, and their coauthors
describe the field as both admirably ambitious and dangerously naive.
But when I look at what those rationalists were after, I don’t blame them for
their naiveté, the way the book’s authors seem to. Instead, I think they were
limited by their time and their tools. What if those same researchers were
recruited into a grand national effort today? They wouldn’t be working for
RAND, or even for DARPA. They’d be working at Facebook, or Google, or
Amazon. And while the rudimentary flowcharts and columns they used to apply
theoretical systems like the Prisoner’s Dilemma look facile today, those were the
same instincts—to find algorithms that could capture, predict, and even shape
human behavior—that created AI, social media, and Big Tech. They just didn’t
have the horsepower or the fuel supply we do today.
“No rule, whether of logic or probability theory or rational choice theory,
mechanically applied, is likely to be able to handle the gamut of political
choices,” Erickson and coauthors wrote with some frustration, and on this we
agree. But time and again, we humans fall in love with logical-looking systems
the same way Cold War rationalists did. We can’t resist the temptation of
offloading hard decision-making, and it’s happening now more than ever. As
we’ll discuss further, researchers inside for-profit companies today are deploying
the same reductionist instincts, along with huge amounts of computing power and
algorithmic sophistication, to study what we choose to do in our lives. And while
the stakes may not be as high as nuclear warfare, they are still frighteningly high.
Because while the mind reels from the threat of annihilation by world-destroying
weapons, it can’t access its instantaneous snake-detecting, action-taking System 1
when faced with far-off threats. Even climate change, which threatens us as a
species, isn’t something we seem to feel until we’re directly faced with flood or
fire. And AI is a far subtler threat than that. In fact, for all the reasons we’ve
discussed, we’re primed to accept what AI tells us. And that acceptance is already
getting us in trouble.
JUST AFTER 5 p.m. on April 9, 2017, passengers on United Airlines flight 3411
sat packed together at the gate at Chicago O’Hare, waiting for the plane to push
back and begin its journey to Louisville. But the plane didn’t move. Instead, a
voice came over the cabin speakers to explain the delay. Four crew members
needed to get to Kentucky to staff a flight there. Would any volunteers be willing
to get off this plane in exchange for $400? The next flight wouldn’t be for another
twenty-one hours, though, and with their bags already smashed into place over
their heads, the passengers were in no mood for compromise, and ignored the
plea. The offer was raised to $800. No takers. And so the voice returned, this time
announcing that four names would be selected, and those people would have to
leave. A moment later, flight attendants approached a young couple who dutifully
rose from their seats and trudged forward off the plane. But when the attendants
approached Dr. David Dao, he didn’t rise. Approached by a United supervisor,
Daniele Hill, Dao identified himself as a physician on his way home, and he
pointed out that missing tonight’s flight would mean missing his rounds in the
morning. He had patients depending on him to be there. He could not get off this
plane.
Tensions rose. Hill was adamant that Dao had to leave. Dao was adamant that
he could not. According to other passengers, Hill told Dao, “I’ll have to call the
police and have you escorted off the plane.” And that’s what happened.
As a hub of two major airlines, Chicago O’Hare moved nearly 80 million
passengers that year, and the incredible volume of humanity that passes through
the airport means not only that it’s run by its own city agency, the Chicago
Aviation Department, but that the department has its own police force. The
roughly three hundred officers of the Chicago Aviation Police were never
supposed to be full-fledged police: forbidden to carry firearms, they had only
limited authority to make arrests, and in those cases their training was to hand
suspects off to Chicago PD. But they were trained like police and had to meet
roughly the same physical and educational requirements.
That night in April, Chicago Aviation Police officers James Long, Mauricio
Rodriguez Jr., and Steven Smith walked down the aisle and told Dao to get up.
Seated in the place he’d paid for, Dao refused. And then, tearing through the
fabric of corporate-customer relations we all assume protects us, they put hands
on him. Gripping him as he struggled, Long, according to another passenger,
yanked Dao sideways, and the doctor’s head bounced off the metal armrest. He
went limp, and they dragged him by his arms past hundreds of gaping passengers.
Phone video of the incident showed Dao visibly bloody and unresponsive. The
United crew needing to reach Louisville got on the flight, and it departed, arriving
two hours late.
The next day, as footage of the incident began to circulate on social media,
United CEO Oscar Munoz put out what is today considered a famously disastrous
piece of corporate communications, claiming in a statement that Dao’s removal
was necessary for “re-accommodating the passengers.” An internal memo to
United staff criticized Dao and praised the actions of Hill and her team. But
within a week, as heat built on social media and United’s share price began to fall,
Munoz was apologizing to anyone who would listen. “No one should ever be
mistreated this way,” he told one interviewer.
Two weeks after the incident, on April 24, Dao announced through his lawyers
that he would sue, claiming a broken nose, missing teeth, a concussion, and
damage to his sinuses that required surgery. Three days after that, his lawyers
reached an undisclosed settlement with the airline.
The story, in its endless retelling, has become a parable about bad choices
under stress—bad choices that radiated upward from the United supervisor Hill to
the CAP officers to United’s CEO. But why did they make those choices? I
believe they were caught in the loop. Not just the first loop, the species-wide one
made of our most ancient instincts. The second loop, where those ancient instincts
are played upon by new, invented systems.
After the passengers had steadfastly ignored several offers of money and hotel
rooms, when the United staff could not cajole anyone to give up their seats, as the
crew for the unstaffed flight waited with their bags at the gate, it was not at all
clear how to break the stalemate. So a nonhuman system was invoked. Names
will simply be chosen, the staff told the passengers. And suddenly everything
changed.
Until that point, the situation was a human-to-human, transactional one, in
which one thing of value (a voucher for a free ticket) was being offered in place
of another thing of value (the chance to fly home and sleep in one’s own bed).
The payment from the passengers would be their inconvenience. As a transaction,
this portion was transparent: why the deal needed to be made, what was on offer,
what the cost would be. But the company wouldn’t pay enough to make the
voucher-for-overnight-stay mechanism work. With seemingly no other options,
United’s people on the ground ended the transparent human negotiation and
activated an opaque system by which an unseen machine would simply choose
which passengers had to get off the plane.
It turns out that when humans are subjected to a system they don’t understand,
they abandon many of the critical faculties they otherwise bring to bear in a more
human interaction. There is an enormous amount of scholarship about this habit
of mind, stemming mostly from anthropomorphism, the tendency to credit a
confusing system with sophisticated attributes it does not possess.2 Carey
Morewedge, a professor of marketing at Boston University who studies decision-
making and bias, told me, “Our findings suggest that when the domain has
uncertainty, people tend to ascribe intelligence to explain outcomes in those
domains.” In this case, the passengers on the airplane, informed that names would
be simply chosen, were suddenly at the mercy of a system they didn’t understand,
and their anthropomorphism was activated. If the airline chose those people to get
off, passengers presumably reasoned, they must be the people who have to get
off.
And other researchers have found that System 2, the one Kahneman and many
others have shown is a delicate, inefficient, but extremely important error-
correction system for our automatic System 1, is also easily distracted by
nonsensical but convenient assumptions. The psychologist and University of
Chicago professor Jane Risen studies decision-making through topics like
superstition, going with one’s gut, and other forms of magical thinking. In a 2016
study, she found that these tendencies are often so compelling that “people cannot
seem to shake them, despite knowing that they are incorrect. In these cases,
people successfully detect an error of rationality, but choose not to correct it.”
Sports fans rationally know they have no role in the outcome of their team, no
matter what they do in their living room at home, she writes. Half of Americans
know there’s no rational reason to knock on wood for good luck, but nonetheless
admit they regularly do so. Our rational mind hands over the steering wheel:
“System 2 acquiesces to a powerful intuition. Thus, in addition to being lazy and
inattentive, System 2 is sometimes a bit of a pushover.”3 When an outside system
is entirely opaque, and also plays into our magical thinking, as demonstrated by
the passengers and crew and airport police on the United flight, everyone except
David Dao, we do indeed get pushed over. Even in the face of patients needing a
doctor or a man violently dragged down an airplane aisle, we find a way to trust
the logical-looking system.
Weizenbaum wrote later that he built ELIZA to learn what he could about
natural-language processing, but wound up discovering far more about human
tendencies. And in the end, those tendencies caused him to flee the field. He spent
the rest of his life critiquing the sort of work he’d been doing, and he died in 2008
after a long period of environmental activism. In his 1976 book, Computer Power
and Human Reason, he described a series of shocks that had him rethinking his
project and recoiling from its implications. Shortly after creating the first
functional prototype, for instance, he recalled asking his secretary at MIT to test it
out. She had observed him working on it for months, and presumably knew it was
a computer program. And yet “after only a few interchanges with it,” he wrote,
“she asked me to leave the room.” The conversation had become too intimate for
her to let her boss eavesdrop. Weizenbaum knew that people form relationships
with tools, cars, all manner of machines, but “what I had not realized is that
extremely short exposures to a relatively simple computer program could induce
powerful delusional thinking in quite normal people.”
And that delusional thinking went far beyond Weizenbaum and his secretary.
Weizenbaum discovered that experts in human behavior were only too interested
in the possibility of robot-enhanced therapy. (In fact, I have several times
described to business-school and engineering students Weizenbaum’s ELIZA
work and the ethical dilemma it posed for him. The reaction is largely shock that
he walked away from so promising a business model.) The Stanford psychiatrist
Kenneth Colby wrote about using something like ELIZA to supplement the job of
human mental-health workers. “If the method proves beneficial,” Colby wrote in
a 1966 paper, “then it would provide a therapeutic tool which can be made widely
available to mental hospitals and psychiatric centers suffering a shortage of
therapists.”5 In the 1970s, Colby expanded on the ELIZA concept by building
PARRY, a software simulation of a paranoid patient (used to train student
therapists) that was indistinguishable from a human patient to most psychiatrists
—it was the first piece of software to pass the Turing Test, which evaluates
whether a person can tell the difference between a robot and a human in
structured conversation. By the 1980s, Colby had sold a natural-language
psychotherapy program called Overcoming Depression to the Department of
Veterans Affairs, which distributed it to patients who then used it without direct
supervision from a therapist. Weizenbaum was alarmed to discover that even Carl
Sagan, the most visible spokesperson for the wonders of science, was really into
the idea. “I can imagine the development of a network of computer
psychotherapeutic terminals,” Sagan wrote about ELIZA in 1975, “something like
arrays of large telephone booths, in which, for a few dollars a session, we would
be able to talk with an attentive, tested, and largely non-directive
psychotherapist.”6
Computer Power and Human Reason was the beginning of Weizenbaum’s
new public life as a critic of artificial intelligence, and, more broadly, of our
willingness to do what it tells us. In a 1985 interview with MIT’s The Tech
newspaper, Weizenbaum described the whole concept of computing as “a solution
looking for a problem.”
People come to MIT and to other places, people from all sorts of
establishments—the medical establishment, the legal establishment, the
education establishment, and in effect they say, “You have there a very
wonderful instrument which solves a lot of problems. Surely there must be
problems in my establishment—in this case, the educational establishment,
for which your wonderful instrument is a solution. Please tell me for what
problems your wonderful instrument is a solution.”7
WHAT AI ISN’T
EVEN THE PHRASE is trouble. From the first time humans paired the words
“artificial” and “intelligence,” we were playing with fire.
That first time was 1955, when a young assistant professor of mathematics at
Dartmouth College, John McCarthy, tired of the disorganized overlap of nascent
fields like “cybernetics” and “complex information processing” and “information
theory,” decided he wanted to bring colleagues together for the summer to try to
create a little clarity around the concept of thinking machines. His funding
application, which he coauthored with Harvard’s Marvin Minsky, IBM’s
Nathaniel Rochester, and Bell Labs’ Claude E. Shannon, was titled “A Proposal
for the Dartmouth Summer Research Project on Artificial Intelligence.” It’s
considered the first time the phrase was ever printed. The idea, as McCarthy and
his coauthors put it, was to put a group together and “proceed on the basis of the
conjecture that every aspect of learning or any other feature of intelligence can in
principle be so precisely described that a machine can be made to simulate it.”
Ten invitees attended some portion of eight weeks across the summer of 1956,
paid for with $7,500 from the Rockefeller Foundation, and some four dozen
people in all wandered in and out at some point, crashing in various empty
apartments as well as the Hanover Inn, and generally freaking out together.
Thirty-year-old Ray Solomonoff, newly minted by the University of Chicago
with a master’s in physics, was the only attendee, aside from cohosts McCarthy
and Minsky, to spend the whole summer at Dartmouth. He seems to have been the
hype man of the group. He took extensive longhand notes on the gathering,
embodying in his loopy logic and frenetic scribbles the communal frenzy of the
experience. He even began to tinker with his personal food-and-sleep schedule so
he could cram it all in.
“Try eating twice a day: noon and midnite [sic],” Solomonoff scrawls at one
point, between staccato notes about predictions and optimum nonlinear filters.
“Sleep 1am–9am. Only trouble is eating with others occasionally, but I think I
could manage this.… The big disadvantage here may be not taking advantage of
early waking acuity for problems worked on just before sleep.”
This manic burst of creativity led to breakthrough notions, like the idea of a
computer that can solve a problem it has never seen before by drawing on past
experiences with other problems. It was a revolutionary idea. As Solomonoff
wrote, a program “mite [sic] be given a series of correct examples, then a question
without the correct answer.” To make such a machine meant going against the
traditional idea of a computer using only exactly the information presented to it.
Perhaps, he wondered, ad hoc mathematical rules “might do only part of what I
really need” and “non-ad-hoc rules are apt to extrapolate better.” The ultimate
difficulty, Solomonoff wrote, connecting half-finished paragraphs to one another
with long, swooping arrows, was this:
McCarthy and Minsky and Solomonoff and the various thinkers who floated
through the gathering that summer wound up landing on a fundamental concept: a
thinking machine that could take incomplete information and, using rules it had
learned from past and parallel examples, predict that information’s continuation,
whether it be finishing a half-written paragraph of text or calculating the
probability that a past event might occur again. The group was literally ahead of
its time—it would be more than fifty years before what McCarthy and his crew
described was technologically feasible at scale.
The gathering either launched or catalyzed epic careers for practically
everyone who attended. McCarthy went on to develop an “advice taker,” a
computer program that used rules of logic (instead of just raw information) to
solve a problem, as well as LISP, the major programming language of early AI.
He developed notions of time-sharing and utility computing that gave rise to
today’s $250 billion cloud-computing industry. And he later founded and ran
Stanford’s AI lab, while Marvin Minsky ran MIT’s. Wide-eyed, sleep-deprived
Solomonoff went on to propose the first notions of “algorithmic probability” that
could be used for predictions, and created the theoretical framework for using
Bayesian statistics to deal with uncertainty, which makes him the ancestor of
everything from modern weather prediction to AI that can spit out a reasonable-
sounding term paper from a one-sentence prompt. RAND Corporation’s Allen
Newell went on to publish the first doctoral dissertation in AI, “Information
Processing: A New Technique for the Behavioral Sciences.” Prior to the
Dartmouth summer, he and Herbert Simon (an economist and cognitive
psychologist and seemingly the only Dartmouth attendee who wasn’t a computer
scientist) had already built Logic Theorist, the first AI program, and afterward
they used military money to build the General Problem Solver in 1957, a program
that simulated the rules of logic a human might follow.
But let’s cool off from all this praise and excitement. Yes, the ideas were big,
and the work it launched was revolutionary. But that’s not the only reason this
work set off the third loop, the cycle of behavior and pattern recognition and
manipulation that may come to define the next phase of human life. It also set off
the third loop because AI looks impossibly sophisticated and entirely inscrutable.
We, the people whose behavior it will shape, just don’t understand what it is and
what it isn’t. And so it’s important to go back to the beginning of the field to
understand the boundaries of what this group actually gave us. We need to trace
the thread of their work forward from those feverish midnight conversations in
1956 to what we now call artificial intelligence today.
The full history of AI and its development has been documented by writers
more thorough and more talented than I am, and I won’t attempt to summarize the
whole thing here, but it’s worth understanding that the field fought with itself for
years over whether computers could or even should be taught to handle open-
ended tasks. Some of those objections were ethical, but mostly they had to do
with market opportunity. IBM, where McCarthy went to work after the
Dartmouth summer, was initially angling to be the corporate parent of artificial
intelligence. But in 1960, a corporate study by the company detailed several
reasons not to pursue that strategy, including surveys that found customers were
hostile to—even afraid of—the notion of “thinking computers,” and the company
backed out. Minsky, Newell, and Simon were all prone to grandiose predictions in
which robots would be picking up socks off the floor within a couple of decades.
When that didn’t bear out, the disappointment poisoned the reputation of AI. For
nearly ten years beginning in 1974, the whole field’s funding essentially dried up,
the first so-called AI winter. And during the decades-long hunt for better systems,
the field lost its way more than once. Minsky, for instance, published a whole
book criticizing the idea of neural networks—interconnected systems of thinking
“neurons” that might coordinate to attack a complex problem—and the field
largely abandoned the concept for a decade. Neural networks are now a crucial
part of modern AI.
And then there were the fundamental limitations of the era: a lack of
computing power to crunch numbers, insufficient database capacity for the
awesome amounts of information necessary to train an algorithm, and the fact that
those limitations made it impossible to scale up small, one-off experiments into
useful real-world systems. It took two funding winters, countless research dead
ends, and exponentially greater computing power and data-storage capabilities to
arrive at the present moment. Today, various flavors of machine learning, from
deep-learning neural networks to the generative adversarial networks that pit two
neural nets against one another, can do everything from read a printed menu to
steer a car along a winding mountain road.
Again, this all sounds very hot—and it is amazing stuff. But what is actually
being delivered into your life needs to be understood clearly, so we can see what
it does and, more important, what it doesn’t do.
First, artificial intelligence is not robotic intelligence poised to replace the
many talents of the human mind. Before we get excited about that concept—a
general artificial intelligence with the smooth-talking, all-purpose utility of
Hasselhoff’s K.I.T.T. car or Skynet—let’s be clear that despite a famous survey of
AI experts in 2013 showing that some 50 percent of them thought we’d have GAI
between 2040 and 2050, (A) it doesn’t exist at the moment, and (B) as we’ll get
into later in the book, all-purpose intelligence may not even have a market if
simpler forms of AI take over our processes the way they likely will.
So, to pose a question that inspires even the most mild-mannered computer
scientists to rain punches on each other, what is AI? Fundamentally, artificial
intelligence is any system that can be assigned a task, learn from data, and adapt
along the way. But there are, of course, many ways to accomplish this. I’ll try to
summarize a few flavors of it to give a sense of what we mean when we talk
about decision technology, which will inevitably earn me a shouting at from
experts at every event I attend from now on.
Machine learning refers to algorithms that get better at a task through
experience. Machine learning draws on past patterns to make future predictions.
But it cannot reach out beyond the data it has; to make new predictions, it needs
new data. There are several forms of machine learning in common use at the
moment.
First, supervised learning refers to systems shown enough labeled data and
enough correct answers (“this is an orange; this is an orange that has gone bad;
this is an orange that is ripe and healthy”) that it can pick out patterns in the data.
Ask it to identify specific outcomes (a ripe orange, an orange that will be ripe
after a week of shipping, a rotten orange), and if it has seen enough of the patterns
that correlate to those outcomes in the past, it can spot the patterns that will likely
correlate to the same outcomes in the future.
Second, unsupervised learning refers to systems that are given data without
any guidance—no right or wrong answers, no helpful labels, nothing—and
attempt to sort them, typically into clusters, any way they can. Whatever common
patterns differentiate the clusters become the sorting mechanism.
Third, reinforcement learning is another way of processing raw, unlabeled
data, this time through reward and punishment. A training algorithm infers what
you want out of the data, then flogs the system for incorrect answers and rewards
it for correct answers. In this way, reinforcement learning teaches the system to
sort out the most efficient means of avoiding punishment and earning praise. With
no labeled data, it just goes on a craven search for whatever patterns are most
likely to earn a reward.
Let’s apply any or all of these three flavors of machine learning to a single
task: distinguishing cows from dogs. Imagine we’re in a theater. Ranged across
the stage are a dozen dogs and a dozen cows. Some of the dogs are sitting, some
are standing, but it’s hot in the vast room, so all of them are panting. Among those
dogs wander the poor, bewildered cows, complaining with anxious moos about
the heat, the uncertain footing at the edge of the stage, the lack of grass and water.
Now we’ll ask our machine-learning system to tell us: which animals on this
stage are cows, and which are dogs?
The trouble, as humans, is that we’re going to want to believe whatever the
system winds up telling us. System 2, our creative and rational mind, as we
learned in the first chapters of this book, is only too happy to hand over decision-
making to System 1, to our instincts, to our emotions, to any credible system that
has a ready answer. And so it’s important to prepare ourselves for just how tidily
the algorithm will, eventually, deliver a reliable performance in differentiating the
canines from the bovines. It’s going to look extremely credible to System 1, and
thus it will also look credible to System 2. And in a sense, it is credible. Feed
photos or other data about what’s on the stage into the system, and—bing!—out
will pop a list, or captioned photos, of dogs and cows. With enough time to refine
its process, the system will, eventually, astound us with its gift for telling us that
this panting poodle is a dog, and that this heifer nosing the curtains is a cow. But
how did the system get there? It matters because, as we’ll see, this technology
goes about answering stupid questions and world-changing questions the same
way.
A data scientist handed this task would want to know more about the goal of
the project in order to employ the most suitable flavor of machine learning to
accomplish it. If you wanted the system to identify the cows and dogs in a
photograph of the stage, for instance, you’d probably feed it into a convolutional
neural network, a popular means of recognizing objects in a photograph these
days. The system would, by the time you used it, have been trained on thousands
of photographs of dogs and cows (supervised learning!), passing each photograph
through a convolutional layer, which scans and simplifies each image, and then
through a max pooling layer, which breaks the image into sections, and keeps
only the section with the most representative information, before identifying the
image as a dog or cow. (And in order to get good at the distinction, that
supervised-learning process would have first been endlessly adjusted by human
“trainers” who marked the results correct or incorrect, typically paid pennies per
photo on a site like Amazon’s Mechanical Turk.)
If, instead, the system was only given a list of attributes—size of animal,
color, panting or not, hooved or soft-footed—reinforcement learning might
simply begin guessing (dog! cow!) and adjust its verdict as the training algorithm
compares those guesses to what it infers the humans want out of the matter. That
process might take hours, days, maybe longer, but if no one had bothered to build
a dog-versus-cow recognition system before, perhaps it would be the best route to
distinguishing the animals from one another.
Whatever AI we brought in, there are two major forces at work that guide the
outcome of the process. The first is called the objective function. It’s what the
human wants out of the project, whether that’s a self-driving car parked
equidistant between two other cars, no more than six inches from the curb, or a
cheeseburger parked between two buns, cooked medium rare. The objective
function is the purpose toward which the whole system is striving, and writing it
out clearly is the defining first task of any successful machine-learning system.
Ancient myths are full of badly composed objective functions: Pirithous wants
Persephone, queen of the underworld, as his wife and winds up stuck in hell.
Sibyl asks Apollo for as many years of life as grains of sand in her fist, but she
forgets to ask for youth along with it and winds up shriveled in pain. If the self-
driving car isn’t told it must park parallel to the curb, or if the burger bot doesn’t
know the buns go on the outside, it can all go very wrong.
The other major force at work is the sheer ruthless efficiency that we design
any machine-learning system to use in pursuit of the objective function. After all,
the purpose of these systems is to save humans the time and effort required to
spot patterns in a vast field of data, and to get better and better at the process over
time. That means machine-learning systems working on the difference between
cows and dogs are going to latch on to the first reliable means of differentiating
the two, whether that’s some combination of size and slobber (the dogs are
smaller and, because they’re panting in the heat, they’re more slobbery; the cows
are bigger, and while they’re also suffering from the heat, they can sweat as well
as pant) or the way the silhouette of the cows is more sharply defined against the
curtain than the silhouette of the dogs, which happen to be fluffier.
Even if the two primary requirements are met—a clearly predefined objective
function and a ruthlessness with which the system will pursue it—it will be hard,
perhaps impossible, for you or me, waiting impatiently for our list of cows and
dogs, to know that that’s the case. In fact, it will be largely impossible for you and
me to know anything about how the process works. Because although even if in
the end our stage full of dogs and cows will eventually be categorized properly,
the process by which our machine-learning system does the job will, by and large,
be invisible to us. All we’ll know is that we put an objective function into one end
of the system, and after some trial and error we got a reliable-looking answer out
of the other end. What happened in between, even for the specialists who built the
system we used, is going to be largely a mystery.
As a result, there is, in fact, a movement in artificial intelligence referred to as
“explainability,” a push for transparency that is both ethically important and
technologically difficult. Because while McCarthy and his crew would rightly be
thrilled by what has been built since 1956, they surely didn’t understand that
extremely impactful decisions would be made by the sorts of automated systems
they envisioned—who gets a loan, a job offer, the chance to get out on bail—
without the ability to see how, exactly, those systems arrived at their conclusions.
This is referred to as the “black box” problem. Modern machine learning gives us
answers but doesn’t show its work.
A 2018 competition called the Explainable Machine Learning Challenge,
organized by Google, the Fair Isaac Corporation (FICO), and five universities,
sought to force a solution. The FICO score is used in the United States to
determine an individual’s creditworthiness, based on an average score from the
three main credit-reporting companies. (It is itself an infamous black box, in
which Americans’ ability to do crucial things like rent or buy a home is
determined by a three-digit number into which they have almost no insight, and
little recourse.) FICO provided the competitors with a data set of anonymized
individual credit history information, along with whether or not each person had
defaulted on a home loan. The challenge then asked the participants to build a
black box system that would predict whether a loan applicant would go on to
repay the debt, and to then explain how the black box arrived at its conclusions.
When deployed on big, unwieldy problems, every flavor of artificial
intelligence tends to put together enormously complicated connections between
the variables they are handed. For a human to identify how a system puts those
variables together is itself as complicated as anything the system may have been
designed to do. And that means that from the earliest prototypes, modern
machine-learning systems have been more or less allowed to simply do what they
do without much in the way of inspection. In fact, it’s become a piece of accepted
wisdom in machine learning over the years that if a system has to be forced to
show its work along the way, it won’t be as sophisticated or as accurate in what it
can do.
This might not matter in small tasks like a cow-versus-dog test. But as
machine learning begins to use human behavior as its data set, and predictions
about what art we’ll enjoy, what jobs we’ll do well, what crimes we’ll commit
become its objective function, understanding the interstices of the system may be
a matter of moral and legal necessity. At the moment it is not, largely because
building a system that explains itself is simply harder to do than building one that
doesn’t.
Currently, building what’s called an interpretable piece of AI involves
observing certain constraints along the way. The designers might limit the number
of variables under consideration so that they will know later that the system relied
on at least one of a finite number of data points in reaching its conclusion. (The
coat of the cows and dogs, their size, their shininess, their gait, but nothing other
than those things, for instance.) The designers might combine smaller AI models
that are themselves understood so that inspectors can decompose the system into
its constituent parts down the line.
But it’s not clear how one can reverse engineer the decision-making process of
AI that hasn’t been designed for transparency. The participants in the Explainable
Machine Learning Challenge were being asked to build something intentionally
opaque, feel around inside it for some sense of how it does what it does, and then
try to turn that into a meaningful system of explanation. It was a horrifically
difficult task.
And a six-person team at Duke University took a look at this Gordian knot and
decided to cut right through it. They looked at the FICO data and the
requirements of the competition and decided they weren’t going to build a black
box at all. They believed they could build a system that didn’t have to be a black
box. It would issue accurate verdicts the average banking customer could inspect
and understand.
Professor Cynthia Rudin, a computer scientist on the Duke team, and Joanna
Radin, a sociologist and historian at Yale, wrote about the team’s experience in an
article for Harvard Data Science Review:
The model was decomposable into different mini-models, where each one
could be understood on its own. We also created an additional interactive
online visualization tool for lenders and individuals. Playing with the credit
history factors on our website would allow people to understand which
factors were important for loan application decisions. No black box at all.
We knew we probably would not win the competition that way, but there
was a bigger point that we needed to make.2
And they were right: they didn’t win. The organizers of the competition would
not allow the judges to play with and evaluate the Duke team’s visualization tool,
and a trio from IBM Research won the competition for a system based on
Boolean rules that would help a human data scientist inspect the black box. In its
paper describing the winning system, the IBM team rightly pointed out the need
for explainability “as machine learning pushes further into domains such as
medicine, criminal justice, and business, where such models complement human
decision-makers and decisions can have major consequences on human lives.”
That said, their black box would have been just as useful to FICO whether it was
later explained or not. Nothing about its design required FICO or you and me to
understand its inner workings. It just made doing so an option.
But FICO, to its credit, looked at the Duke team’s highly accurate and user-
friendly system, decided that project was worth rewarding, too, and invented a
prize for it for going “above and beyond.”
The problem is, black box systems are already a mainstay of modern business.
Executives who depend on AI are just as oblivious to the mechanism at work
under the hood as you and I would be to how our algorithm told those cows and
dogs apart. FICO went on to conduct a 2021 survey of C-suite financial-services
executives from companies that make more than $100 million in annual revenue.
It found that nearly 70 percent of them could not explain how their AI systems
made decisions or predicted outcomes. And what’s more, by and large they didn’t
seem to care. Only 35 percent of surveyed executives said they’d made any effort
to bring accountability or transparency to the process.
Black boxes are, of course, just easier to build. Wind it up, let it go, don’t
worry about it. But they may have become the norm for a few alarming additional
reasons. First, there’s the assumption that letting a system learn its own rules
behind closed doors is the fastest path to an accurate result. Then there’s the
desire to keep the inner workings of a valuable system from getting out to
competitors, even if that means those who built the system don’t understand the
inner workings, either.
However, there’s also the fact that demonstrating cause and effect is the first
step in any lawsuit against a company deploying a technology like AI. Based on a
1975 federal case, Warth v. Seldin, it has been necessary for plaintiffs to establish
clear causation linking their loss or injury to a defendant for a lawsuit to go
forward. But if plaintiffs can’t even inspect an AI that has denied them a bank
loan, or a job, or a shot at bail, how could they even begin to make that claim?
“To some, the obvious solution to the intent and causation problems,” wrote the
attorney Yavar Bathaee in a 2018 article in the Harvard Journal of Law &
Technology, “will be to either increase the transparency of AI through standards-
based regulation or to impose strict liability.” And yet, he argues, “both
approaches would risk stifling innovation in AI and erecting steep barriers to
entry.” Whether he’s correct on the effect regulation might have, it’s clear that a
company that builds a black box system is currently dodging a whole host of legal
and public-relations headaches. It may be that we’ve come to consider black
boxes not only more accurate and easier to protect from competitors, but that
we’ve also adopted them as a bulwark against legal exposure.
The black box paradigm “allows the model creators to profit without
considering harmful consequences to the affected individuals,” Rudin and Radin
wrote. “It is possible that an interpretable model can always be constructed—we
just have not been trying.”
There are companies that have tried. Verikai is an AI-driven risk-assessment
company that tries to give insurers the ability, through machine learning, to
underwrite health insurance for companies smaller than five hundred people.
Doing so has been very difficult in the past, because the law of big numbers in
health insurance meant needing the largest possible group of employees in order
to accurately predict how many of them in a given year might develop diabetes,
cancer, or other diseases. Below five hundred people the predictive power begins
to fade, and below two hundred fifty people it tends to fall apart completely.
Verikai has instead sought to extrapolate risk by using AI to spot patterns in
enormous numbers of people, and then carrying those predictions into small pools
of employees. Using data brokers like Acxiom and Experian, among many other
sources, “we are scoring people’s sleep patterns, alcohol consumption patterns,
eating patterns,” Verikai’s former president, Chris Chen, told me. AI is essentially
allowing Verikai to draw conclusions about what humans do in general, and then
carry those conclusions into smaller representative groups.
People who report better amounts of sleep, or who consistently adhere to
taking their medication, correlate to a better chance they’ll repay their mortgages,
Verikai has found. But “there’s other really random correlations,” Chen told me.
And this is where AI is not only making deeply influential recommendations—
recommendations that will affect fundamental aspects of your life, like whether
your employer decides it can afford to help you pay for health insurance—without
showing its work. It’s also making deeply influential recommendations based on
stuff you didn’t even realize it was using to reach conclusions about you. “You
own a swimming pool and you have, say, these three other attributes and as a
result, you’re lower risk and we’re not sure why,” Chen offered as an example.
But he also doesn’t believe companies need to blindly accept those correlations.
“Everyone talks about AI needing to be a black box,” he said. “It really doesn’t
need to be.” Instead, Verikai’s system is built to actually show the public what
aspects of their lives suggest a certain degree of health risk, according to the
patterns the AI has spotted. The company’s AI draws correlations between
seemingly unrelated predictors of your health outcomes, but at the very least the
company can show you what they are. Most won’t do you that courtesy, and
there’s little incentive for them to start. By and large, we have no idea how these
technologies are working on us; we just assume they work.
FOR ROUGHLY THREE hundred years, one irrational theory dominated the
thinking of rational people trying to explain the universe. It was a theory that
sprang from ancient superstitions, and for which there was no measurable
evidence, but it was so useful in its simplicity that it became an article of faith to
which generations of researchers remained committed even into the twentieth
century.
It was the idea of “the aether,” an absolute stillness in the universe against
which movement—like that of light—can be measured.
Plato and Aristotle got it started by referring to a “fifth element” (quinta
essentia) that sat in some way above Earth, Air, Fire, and Water. (The ridiculous
and wonderful 1997 film The Fifth Element revolves around the idea that the
quinta essentia is love, but that’s not what we’re talking about here.) The
philosophers each theorized that this quintessence made up everything beyond our
sky. And while most ancient Greek ideas about it were soon dismissed—that it
moved in a circular pattern, for instance, while the other elements somehow
traveled along linear paths—the idea stuck around that there was a “pure” element
in the universe, a clear and untainted medium through which things move.
Early medieval alchemists were fixated on the idea that all things had a
quintessence—a purest form—and that through the proper chemical processes
they could distill it from various substances and use it as a curative. The
fourteenth-century alchemist John of Rupescissa popularized the idea that an
alcohol-based quintessence could heal affliction and forestall death (an attractive
notion at a time when most people, including Rupescissa, died before the age of
fifty).3 Alchemists tended to defend their work with mystical explanations,
however, and by the end of the medieval era, as the Renaissance ushered in a new
era of evidence-based science, the notion of quintessence as a magical medication
had largely faded away.
But the idea of a pure essence lived on, and in the seventeenth century René
Descartes began to formalize it in the realm of physics. He was trying to move
beyond what he considered outdated medieval theories when he used the term
“aether” to describe a great ocean in which everything swims: in his view, the
aether was a theoretical medium across which distant objects were still physically
connected, and through which something like magnetism “moves.” This notion
helped to create a mechanical explanation for the behavior of light, and physicists
continued to depend on and refine it. While Descartes considered the aether to
have a static property, Robert Hooke theorized that it vibrated, Christiaan
Huygens theorized that it consisted of spinning particles, and then Isaac Newton
wiped them both out with his suggestion, delivered in the same memorandum to
the Royal Society in which he explained gravity, that “all space is occupied by an
elastic medium, or aether, which is capable of propagating vibrations in the same
way that air propagates the vibrations of sound, but with far greater velocity.”4 In
Newton’s view, small, rapidly moving corpuscles made up the aether. His theory
stood for at least a century.
Newton’s idea lasted so long perhaps because it was so handy, a point of
reference that made calculations easier and theories more convincing. And by the
nineteenth century, the existence of an aether was not only assumed to be a fact, it
had been built into other theories about the sprawling mechanisms of nature. An
1889 correspondence between two pioneers of electromagnetism, Oliver
Heaviside and Heinrich Rudolf Hertz, shows just how deep this faith ran. “We
know that there is an ether,” Heaviside wrote to Hertz with a certainty typical of
the era, while acknowledging with frustration that it had been unobservable. “As
for the structure of the ether itself, that is a far more difficult matter, and one that
can never, it seems to me, be answered otherwise than speculatively.”5
But then a new generation of nineteenth-century physicists began to squint at
the whole notion of the aether, because certain aspects of the theory just weren’t
working out. For one thing, it wasn’t clear why something like our planet could
move through space without any obvious sign of being slowed by the friction an
aether should theoretically impose. Instead, the Earth travels around the sun at a
steady 67,000 miles per hour. And if it were moving across an aether, that aether
should give off some sort of wind, and Earth’s movement should be affected
depending on whether the Earth was traveling with that wind or against it. And as
these granular questions became a priority, cutting-edge experiments began to
reveal that generations of scientists had been entirely, embarrassingly wrong.
Between April and July 1887, a pair of American physicists, Albert Michelson
and Edward Morley, conducted experiments on the outskirts of Cleveland at what
is now Case Western Reserve University. Michelson had prototyped the
interferometer they’d be using for the work while doing research for the navy, and
spent most of the prior few years laboring so relentlessly toward the goal of
detecting aether wind that he’d had a nervous breakdown in 1885. When he and
Morley finally settled into their spring and summer together in 1887, safe from
interference in the basement of a stone dormitory, they had very high hopes
they’d built a device sensitive enough to detect the difference between light
traveling with the aether wind, and light traveling against it. And they had. But
when their interferometer split light from an oil lamp, sent each beam at
perpendicular angles up and down the arms of the device, and then brought them
back together again, the beams arrived at the same time. There was none of the
delay they’d expected. “The Experiments on the relative motion of the earth and
ether have been completed and the result decidedly negative,” Michelson wrote to
a British counterpart. And in experiments by other physicists in 1902, 1903, and
1904, the same results came back. Suddenly it seemed that the all-important and
all-convenient aether was not a thing. It didn’t exist.
With the idea of a constant medium no longer workable, old theories
collapsed, new theories were necessary, and a physicist born in Germany just a
few years before Michelson and Morley emerged disappointed from their
basement lab came up with the best theory around. In 1931, long after he’d
achieved fame for his scientific achievements, Albert Einstein met Albert
Michelson for the first and only time at a dinner in Einstein’s honor at Caltech.
According to his biographer Albrecht Fölsing, Einstein gave an after-dinner
speech for a crowd of two hundred that included the aging physicist, and thanked
Michelson directly for the work he had done when Einstein “was still a little boy,
hardly three feet high.”
“It was you,” Einstein said to Michelson, who died just four months later,
“who led physicists onto new paths and by your wonderful experimental work
even then prepared the road for the development of the relativity theory.”6
What Albert Einstein established from the ashes of the aether was that the
universe was a whole lot more complicated than Plato, Aristotle, Descartes,
Newton, or anyone else had understood. And yet he showed we could measure it.
We just couldn’t measure all aspects of it at the same time. His theory of relativity
had shown, for instance, that space and time couldn’t be measured separately
from one another—they were connected. And because of the space-time
continuum, and the effect that massive objects have on it, events that take place at
the same time for you can take place at different times for me. In fact, Einstein
had found, to gather the information necessary to identify one fact about an object
—like where it’s headed in space—an observer would have to sacrifice the
information necessary to identify another fact—like its position at that particular
moment. There was no stillness. There was no ocean across which things moved
and against which they could be measured. Everything is moving in relation to
everything else all the time, and our math had to adapt to this new fluid model.
I believe we are still in a phase of decision technology in which we naively
consider there to be some sort of simple aether against which we can measure the
effectiveness of what we’ve built. And it’s called an objective function.
In a post on the popular AI blog Abacus, Daniel Kronovet describes the notion
of an objective function in a way that Descartes, Newton, and other pre-relativity
physicists would have recognized:
Put metaphorically, we can think of the model parameters as a ship in the
sea. The goal of the algorithm designer is to navigate the space of possible
values as efficiently as possible to guide the model to the optimal location.
For some models, the navigation is very precise. We can imagine this as
a boat on a clear night, navigating by stars. For others yet, the ship is stuck
in a fog, able to make small jumps without reference to a greater plan.7
But AI, especially when it’s called upon to parse the patterns in human
behavior, is navigating something much more complicated than water, toward
something much more complicated than the closest destination, and the objective
function for you may be very different from the objective function for me.
I once sat in a meeting where a group of researchers put their prototype for a
piece of “ethical AI”—a system one could trust to take ethics into account as it
did its work—in front of a roomful of behavioral and political scientists. (I
attended under an agreement that I could describe the meeting’s content, but not
its participants.) The lead researcher, clicking through the slides behind him,
described his team’s plan: Build a long list of half sentences along the lines of
“As a coworker, it would never be appropriate to [blank].” Have humans
complete a couple thousand of those sentences. Then train AI to spot whatever
hidden patterns there are in the result, and begin filling in the blanks itself. Given
enough training data, he concluded, putting the remote down on the desk, “we
think we can arrive at a set of universal human values. Thank you. And now I’ll
take your questions.”
Every hand in the room went up. A political scientist went first. “I guess I
have three questions for you,” she said. “First, what is ‘universal’? Second, what
is ‘human’? And three, what are ‘values’?” And with that the meeting imploded.
Social scientists, without using the term, have been squinting at whether our
species has an “objective function” for more than a century. Kahneman and
Tversky and Thaler and Banaji and all of them have looked at the fundamental
programming that makes us who we are, and for the adjustments that can be made
to make us better. But increasingly, a new crop of social and political scientists
are beginning to step back from the problem and ask whether in some cases—or
even most cases—there’s any universal notion of “better” we could ever agree on.
They’re pointing out that better for you or me could be in fact far worse for
someone else in other circumstances. And what they’re pursuing suggests that not
only isn’t there a sea across which we’re trying to point a ship, or a destination to
which we can navigate, but also that it’s in fact dangerous to build an automated
system that measures success against a universal medium, in service of a
universal goal.
At this point it’s probably clear that I’m laying a great deal of my concerns at
the foot of AI itself—its craven devotion to pattern recognition, our overinflated
opinion of it, the way it hides its work. But I don’t believe we’ll be able to protect
ourselves from the third loop by only adjusting the way AI is built. We also need
to recognize that the patterns in our behavior don’t always point the way toward a
desirable objective function. Using AI to make important choices in society isn’t
just problematic because of the AI. It’s also problematic because some of our
choices are impossible.
John Patty and Elizabeth Penn taught me this. They’re a husband-and-wife
academic team at Emory University. They’re both professors of political science,
but they’re also mathematicians, and while other political scientists tend to work
in formal theories that treat humans as rational actors with rational goals, these
two use math to try to account for the persistent inability of humans to get along.
In effect, while others are still trying to measure the physics of human choice
against the aether, Patty and Penn have come to accept, just as Einstein did, that
things are much harder to measure than we thought they were.
In their 2014 book, Social Choice and Legitimacy, Patty and Penn write about
whether we can ever bring conflicting individual goals together into a grand
collective decision. And almost immediately it’s clear that these two are trying to
account for the thorniest possible stuff. Because rather than suggest it’s possible,
as civics class always tried to teach me, that people can come to a satisfying
consensus through earnest discussion in a town hall or a national election, Patty
and Penn suggest that satisfaction isn’t the thing we should be after.
“In a nutshell,” they write, “any minimally democratic aggregation procedure
must encounter some situations in which it fails to produce a coherent (or,
perhaps, ‘well-ordered’) collective preference.” As a group, there are some things
on which we simply cannot agree. So what can we hope for? Patty and Penn
argue that rather than just give up on the idea of a popular will, as so many of
their peers have spent careers arguing, we should instead hope to establish a
feeling of legitimacy in how decisions are made, by explaining them and the
principles they’re based on. They know that won’t be satisfying for everyone. It
won’t make for a great speech at the end of the movie. It won’t create a catchy
tune. “Our theory does not provide any particular reason to expect that all
individuals will be equally happy (or, perhaps, happy at all) with the conclusions
reached by such a process. But that is arguably the point of legitimacy: a
legitimate decision is one that makes sense in spite of not being universally
acclaimed.”8
They aren’t writing about AI or objective functions, but they might as well be.
Their theory of social choice, which one reviewer called “a thicker conception of
democracy,”9 suggests that in order to achieve any satisfactory outcomes in law,
politics, or administration, we must feel that the legal, political, and
administrative system knows what it’s doing. And we have to have the principles
that drove a decision explained to us. None of that will be possible if we continue
to go down the road of deploying opaque, proprietary AI systems to evaluate our
world and render verdicts for us.
The expediency of business isn’t just causing companies to ignore the inner
workings and ethical risks of the AI that makes their services possible. In fact,
most companies are reusing one piece of AI over and over again. It is
prohibitively expensive for most companies to purpose-build AI for the task at
hand. Instead, for reasons of cost, they adapt existing algorithms, built around
other objective functions, for tasks in some other domain. And as that has become
more and more the case, the same pieces of machine learning are being deployed
on everything. This creates the risk that certain reused pieces of AI are going to
be applied to tasks for which they aren’t designed, and may carry over
algorithmic habits learned in other data sets that wind up skewing important
results in the one to which it’s been ported. As UC Berkeley computer science
professor Moritz Hardt put it to me, when it comes to developing algorithms that
are fair and effective, “lending is going to be different from criminal justice, and
college admissions are not the same as recidivism prediction.” And when it comes
to getting things wrong, “the cost to society, to the individual, and to the decision-
maker, they’re very different.” Bolting the same forms of AI onto each problem
can save and make money. But those objective functions shouldn’t be considered
universal, he told me. “There’s not going to be an easy, all-purpose definition that
we could just sort of slap onto our algorithms and solve the problem.” And yet in
a world where so few leaders of companies care to even inspect how the AI they
use works, because the technology seems to deliver reliable, efficient results, who
will be on guard against subtle, hidden algorithmic effects and biases that don’t
show up in the bottom line, and may take years to play out?
I wish it were the case that the solution to this problem is simply improving
the algorithms, or requiring explainability, or demanding that companies tell us
when they’ve borrowed AI trained on recipe recommendations to build a system
that sorts through résumés. But it’s not just a design problem.
Once upon a time this sort of world-changing technology was developed by
academic institutions, national labs, the Department of Defense. Today, AI is
being refined entirely inside for-profit companies. It’s built to make money. And
yet the marketing, the libertarian politics, and the continued social acceptability of
technologically driven industries like the one that has produced decision-guidance
technology depends on the assumption that tech makes our lives better. That
assumption drives the adoption of AI, as it drove the adoption of smartphones and
the Internet before that. It’s a reflection of our fundamental Western assumption
that things are always improving, and of our fundamental American assumption
that capitalism creates solutions that will lift up everyone. I worry that in AI
we’ve built something that will so fundamentally reshape human behavior in
pursuit of profit, and will so easily find and amplify its worst aspects along the
way, that all of those assumptions will be proved wrong without our realizing it.
The Unitarian minister Theodore Parker Jr., speaking in 1853 about the
prospects of abolishing slavery, said, “I do not pretend to understand the moral
universe; the arc is a long one, my eye reaches but little ways; I cannot calculate
the curve and complete the figure by the experience of sight; I can divine it by
conscience. And from what I see I am sure it bends towards justice.” In 1950, at
the end of the Montgomery bus boycott, Martin Luther King Jr. condensed the
idea in a speech about the prospect of equality for all Americans. “How long? Not
long, because the arc of the moral universe is long, but it bends toward justice.”
President Obama often cited King’s phrasing in his appearances at the podium.
It’s a beautiful idea, and it may yet prove to be true. But consider that we have
only been thoughtful enough to express that sort of idealistic sentiment for a few
thousand years, and to have ideas like that at all for perhaps 70,000. The history
of humanity simply isn’t long enough to know whether it’s an arc, or which way it
bends, as Parker originally admitted. We’re not on a preordained flight path
toward a better life. It’s not a given that what we build will improve our world.
We are living an improvisation, carefully balancing our institutions, our
partnerships, our social codes on the tip of a collective finger we only just figured
out how to use. What is universal? What is human? What are values?
AT THIS POINT it’s hopefully clear that we’re faced with a conflict between AI,
our ambitions for it, and our tendencies and needs as individuals and as a society.
We give AI an objective function and expect it to reach that goal with ruthless
efficiency—we want an aether against which AI can be measured—but setting
objective functions for humans, given how different we want to be from who we
naturally are, is inherently problematic. But let’s assume, for instance, that the
function of AI should be to make life better for humans in general. How would
we even begin to determine what the objective function of that algorithm should
be?
One AI researcher told me in 2019 that he and his colleagues had been kicking
around a hypothetical ethical dilemma they called “the heroin problem.” (It’s
important to recognize here that the hard, yearslong work that harm-reduction
experts have done to reframe opioid abuse as an affliction, rather than a choice,
makes the binary example he laid out a dated and insensitive one. So also
consider that AI researchers are kicking these ideas around, but that they’re not
often up to date on the latest thinking outside their field.)
He described the heroin problem to me this way: Imagine being asked to build
a deeply personalized AI-based assistant, one that looks for patterns and
automatically offers proactive help, in order to help the customer reach her goals
as efficiently as possible. Let’s imagine that most users adopt the product in order
to help save money for retirement, and as a result the AI begins looking for
signals in the data that correlate to financial growth. The objective function, in
other words, is “save as much money as possible.” (This is not yet a thing, it
should be noted. No one makes money tricking people into not spending theirs.
It’s the reason you’ve never seen a flashy television advertisement for a Roth IRA
or a 401(k).) The AI might instantaneously sock away any extra income in an
investment account, and politely offer to cancel appointments that have been
linked to spending money in the past.
“But now imagine you’re addicted to heroin,” he said to me. “What should the
objective function be?” Should the AI draw its lessons from all the other people it
has analyzed, people who are saving money, and try to nudge me off the habit? Or
should it accommodate my dependence on heroin, by, say, prioritizing time with
other people who use, or shielding me from appointments that past patterns
suggest I’m going to be too sick with withdrawal to attend?
And even if AI could be entrusted to pursue these sorts of goals, how do we
know that the people making it are even interested in pursuing a form of the
technology that can handle the variety and subtlety of human behavior and
satisfaction? AI is being developed in pursuit of profit, not perfection, after all. As
a result, it’s not clear that the handful of companies that dominate the field have
any reason to squint skeptically at the short-term conveniences of AI the way that
nineteenth-century physicists squinted at the short-term conveniences of the
aether theory. Imagine that while Michelson and Morley tuned the arms of their
interferometer in the cool stone basement, researchers elsewhere around the world
had discovered that one could make vast sums of money off aether-based
calculations, whatever the abstract, long-term threats to human understanding of
the universe. If the aether theory had promised to save endless hours of
calculation, and produced results that looked reasonable and produced a return the
way that AI has disrupted industry after industry by matching advertisers to
potential buyers and political opinions to an audience, we’d likely still be talking
about aether winds and a theoretical universal substance. The uses to which AI is
being put aren’t about truth and science and deeper understanding—at least, that’s
not why companies are putting money into developing the technology. Something
like the “heroin problem” is difficult enough on its own, but it foretells disaster
when you consider the profit-making purpose of AI.
And so although it should be clear by this point that using AI to make money
could unleash a decision-making force in society that we’d have no control over,
leading us blindly in all sorts of terrible directions, none of that gets in the way of
the business case for using AI. And that’s where The Loop has begun to encircle
us.
Chapter 8
COLLAPSING SPIRAL
THE FINAL RING of The Loop—the one for which this book is named, the
convergence of pattern-recognition technology and unconscious human behavior
—is still in disparate pieces. It hasn’t fully taken form. But its component parts
are being assembled all over the world, in all sorts of industries, and the way they
can be bolted on to any part of your life (an algorithm that can pick out patterns in
your taste in wine can also credibly pick out patterns in the clothes you might
wear or the live concerts you might enjoy) means they’re going to naturally click
together over time. And if that happens before the behavioral effects become
grounds for legal action or regulation, we’ll see a world in which our choices are
narrowed, human agency is limited, and our worst unconscious impulses
dominate society. Some people seem to believe that AI will free us, that we’ll
wheel through the world unencumbered by busywork or uncertainty. I worry that
as we become caught in a cycle of sampled behavioral data and recommendation,
we will be instead caught in a collapsing spiral of choice, at the bottom of which
we no longer know what we like or how to make choices or how to speak to one
another.
I want to explain in this chapter how easily decision technology spreads into
all corners of our lives. And while the examples I’m about to describe may feel
disconnected, remember that the interoperability of machine learning means a set
of algorithms built to do one thing can also do many others well enough that
you’ll never know its various roles, so anything AI can do in one part of your life
will inevitably metastasize into others. I could start describing this third ring of
The Loop with examples from all sorts of industries. But I want to start with
surveillance, because it’s immediately understandable, it’s already changing our
behavior in unconscious ways, and its effects are rapidly accelerating.
After five weeks of lockdown at the beginning of the pandemic, when it first
became clear that this would be a deep and vast economic crisis, with more than
26 million Americans filing for unemployment by late April, the national and
international mood turned toward a desire for some way to monitor the spread of
the virus while somehow getting back to public life. The president of the Veneto
region in Italy, where the first known Italians had died of the virus, talked about
creating some sort of work-license system for people who had tested positive for
COVID antibodies. Public health officials in the United Kingdom and elsewhere
began openly discussing the possibility of a “passport” given to those with the
necessary antibody test results or a vaccination card that would allow them to
work and travel again. Google and Apple announced a plan to build an app
together that would use Bluetooth to detect other phones nearby, and, if a person
volunteered to have their test results and contacts recorded, it would alert users if
they’d been too close to someone who was later determined to be infectious. The
logic of building systems of detection and certification was obvious. And
determining who was no longer a risk to the public seemed to require some sort of
central public database. The involvement of ubiquitous companies like Google
and Apple seemed inevitable and inarguable. Meanwhile, I began receiving email
after email from companies that claimed to have built contact-tracing systems,
because technologically they were relatively simple to build: any company with a
location-aware app for following the behavior and location and social connection
of its users already had, in theory, the constituent parts to build a contact-tracing
system. This was a rare moment in which building surveillance technology was a
matter of public interest and public health.
Soon a new set of radical surveillance capabilities, largely powered by AI,
became public knowledge. A company called Draganfly, which had invented the
original four-rotor “quadcopter” design for drones, contacted me to describe a
new technology that it said would enable public-health authorities to detect
COVID-19 symptoms from the air.
Researchers at the University of South Australia had already created
technology to rapidly detect, in a single drone pass, the survivors of a mass-
casualty event like an earthquake or tsunami. With 4K-resolution video footage
held extremely still by stabilizers, AI could compare the images in the footage to
training data showing the motion of humans breathing in and out. Once the
system had seen enough footage of people breathing, the researchers found, it
could pick out the living from the dead according to the tiny movements of even
the shallowest, desperate respirations.
Now, they had found that other vital statistics could be gathered in a similar
way. Not only could the system learn to spot breathing, it could subdivide the rate
and depth of the breaths and correlate it to unhealthy types of increased
respiration, and to elevated heart rate. The movement of people coughing into
their elbows could be classified according to the severity of the cough. And the
technology could analyze a person’s skin tone and compare it with people of
similar skin tone to spot people who were especially flushed. And by putting skin
tone and heart rate and respiration together, perhaps the technology could even
spot a fever. Maybe this was the makings of a respiratory-virus detection system.
The researchers had been flying their technology on one of Draganfly’s drones
and contacted the company about a deeper partnership. By April 2020, the
company was ready to fly its expanded vital-statistics program wherever they
could find permission to do so. And in Westport, Connecticut, the police
department said yes.
The Westport Police Department already operates a technology-heavy
surveillance platform, as its chief, Foti Koskinas, explained to me. He’s a rapid
talker and takes obvious pride in his work. Many of Westport’s 28,000 residents
commute into Manhattan, and several financial-services firms have offices in
town, making it a busy train stop. What the roughly 6,000 people getting on and
off a train at the station may not realize is that the police scan their faces as soon
as they hit the platform. “The police department has almost 200 cameras that we
monitor through our train stations and other areas that fall under our jurisdiction,”
he said. “We do have facial recognition on those cameras, we have plate readers.
So we’re doing many things to gather data.” Westport is one of the Eastern
Seaboard’s most prosperous and homogeneous towns, with a median household
income of nearly $200,000 and a population that’s more than 92 percent white.
The accusations of racial bias that have been leveled for years at facial-
recognition systems didn’t seem to have triggered much hesitation on the part of
the police department or city government when it came to deploying it on the
citizens there.
Furthermore, Chief Koskinas said his department has regularly flown four
drones over the town since 2016. “We have so many parks, so many school
campuses, beachfront, an island, we have so many areas that my patrol shifts
cannot possibly adequately cover all of it,” he told me. The drones have been used
to inspect roofs, find missing people, and were already being used to detect and
disperse crowds of people identified by Draganfly’s software as standing too
close to one another in the age of social distancing, when the company proposed
using Westport as a test bed for its symptom-detection software.
As the CEO of Draganfly, Cameron Chell, described it, the purpose was not to
identify any one person in particular. Instead, the first phase of the technology
would be geared toward generating a sense of the rate of infection in a given area.
But soon, Chell told me, he believed the nation would need ongoing
monitoring to get the economy up and running again. “The consensus that we
seem to be getting back is that the new norm is health monitoring on a broad
basis,” he said. Perhaps to reactivate something like sports or professional
gatherings, a drone could tell you what your risks will be before you join a crowd
it has just scanned. “So, you know, do I really want to be entering into convention
centers if they don’t have a health monitoring system that can give us an
indication or a reading of how sick the facility is?”
And that’s what Chief Koskinas wanted to help accomplish. “Today, in
reopening our communities and keeping our communities open, I need every
possible tool,” he told me. “I found it very appropriate for us to explore this.”
And so the week of April 20, an FAA-licensed drone pilot began flying passes
above citizens playing soccer in the park, standing in line outside the local Trader
Joe’s, and walking the streets of downtown Westport, not only to determine
whether they were standing too close (the drone carried loudspeakers so it could
tell people to leave a closed park or to maintain better social distance), but also to
potentially identify symptoms of COVID-19.
Chief Koskinas explained that his department has adopted radical new
technology in the past. Sandy Hook is a half-hour drive north of Westport, and the
2012 murder of twenty children there transformed his department. Before then, he
told me, “we were preparing and training differently. Now it’s common practice
that a regular police officer carries a high-powered rifle every eight-hour shift.”
The drone pilot program seemed like something he had to do. “It might be
something that there’s no future to it. But not looking in that direction to protect
our community? I think that would be irresponsible.”
This particular project didn’t last long. The day I reported on the drones for
NBC News, the police department was so inundated with complaints that it
canceled the partnership. People were deeply disturbed, a Draganfly
representative told me later, to know that they were being recorded, and
evaluated, by cameras flying above them.
But Chief Koskinas said he believes that his department may from now on
have to monitor health as a matter of protecting and serving the residents of
Westport, and so this test was an important part of that.
“So you can imagine a future in which a little league game, a US Open, a
crowded train is scanned by your cameras?” I asked him. “And this technology
would then tell you people are getting sick again?”
“Correct,” he said. Then he paused, realizing that what he was proposing was
a very strange new world. “There’s the side of me, the human side, that looks at it
and says ‘this is very science fiction, or Weird Science’ and it hurts. It really
does.”
We agreed that it felt like the plot of a movie we’d seen before. “But at the
same time, it’s reality,” he said. “It’s our new reality.”
They decided to work with children between the ages of two and a half and
seven, and offered the participants a small stipend. They say they’d hoped they’d
get a few dozen families. Today, hundreds have applied.
Miles sat down in the playroom with his mother, and they leaned together at
the table over an intricate model train set. Domoff, Niec, and I watched from the
observation room, and Niec picked up a microphone and began coaching Heather
in the gentle voice of a children’s book narrator.
Heather had learned from her training here to praise Miles in a specific,
directed, and extremely pleasant way. When Miles picked up the locomotive and
handed it to his mother, his mother kept it simple: “That’s the train engine! How
cool!” We could hear her and Miles on small speakers in our dark room. Niec
immediately piped up. “That’s great,” she says. “Really good labeling, not asking
questions, just showing him you’ve heard him.” And when Heather announced
that soon she and Miles were going to be able to watch a video together on her
phone, and that they’d watch just a little of it, and then go back to playing
together, Niec nodded happily. “Great time management skills there, great job,”
she told Heather.
And then it was time to use a screen together. This is the moment everyone in
the clinic had been anticipating. I thought of my own kids, and how committing
to a video together means rolling the dice on a whole afternoon’s worth of
potential meltdowns. Miles and his mom sat together in silence and watched.
Miles was entirely, unnaturally still. His cowlick never moved. Then Heather
pointed out a detail here, or a character there, and Miles began to narrate along
with her. Over the earpiece, Niec suggested to Heather that she give Miles a
heads-up about winding the video down. I stiffened.
“We’re going to give it another minute, and then we need to move on to
something else,” Heather said, leaning in close. “But you and I are going to get to
play a little more, too, okay?”
Miles nodded.
“Nice job with that preparation,” Niec said. “And see, he accepted that
preparation from you!”
And then Heather asked her son to hand back her phone. There was a moment
of hesitation, and then this kid, who used to scream and break things at moments
like these, gave it to her.
“You did it!” Heather said. “Even though you didn’t want to, but I really
appreciate that you did it.”
Miles leaned over and silently kissed his mom on her shoulder.
I looked at Domoff with my eyes wide. She was already nodding at me.
“They don’t want the screen, they want you,” she whispered.
Heather told me later that this is what they have taken away from the
experience: “Making sure that instead of being like, ‘We’re going to use this
screen time so that you’re good, and I’ll give you time later—’”
“It’s reprioritizing,” Andrew said.
Heather nodded. “By giving our son what he needed first, he’s better able to
manage the transitions.” She thought on it for a moment. “His needs have been
met, and so it does honestly make other times easier.”
Watching this process, I had the clear feeling of discovering a sort of hidden
training that should be available to everyone. Any parent can tell you that being
sent home from the hospital that first day involves the feeling that someone must
have misunderstood, we are in no position to raise a baby by ourselves, are you
kidding me? We have no training!
Here, for me, was the training. Who wouldn’t benefit from having a coach in
their ear during a dozen practice rounds of dealing with a difficult thing my kids
do? And meanwhile there is so much inadvertent, counterproductive training we
go through as parents and children in a world where screens have been
normalized by their constant presence. According to a June 2021 report from
Deloitte, the average US household now has a total of twenty-five connected
devices. Several parents I spoke with told me they get as much guidance on the
proper role of screens from the ads they see by the companies who make them as
they do from anyone with actual expertise into how the technology might affect
their families and their relationships. Only parents who happen to notice the right
flyer in central Michigan, or Seattle, might be getting any professional guidance
about the effects all those screens are having on children (or on parents, for that
matter).
And at the same time, the ancient systems we have for passing information
between parents and children powerfully shape our kids. If parents are infected
with new behavior, it spreads quickly to the children. What are children learning
about the expected role of technology in their lives after a pandemic that had them
watching their parents more or less constantly connected to a device of some sort?
Developmental researchers have long understood that we use our social
sensitivities to offload decision-making from our brains to the brains of others,
and that the offloading from parents to children is especially powerful. In a classic
1960 study, psychologists Eleanor Gibson and Richard Walk built a visual cliff—
in essence, a high box, more than four feet tall. Where it dropped off on one side,
Gibson and Walk laid a thick sheet of glass, capable of supporting a good deal of
weight. They then brought in thirty-six infants between six and fourteen months
old, and had the babies’ mothers call them from the box out across the glass.
The results were an embarrassment for humans. Only nine of the children did
the sensible thing and refused to venture out over the glass. Fully twenty-seven
other children willingly crawled across what looked like a sheer cliff. The
researchers brought in dozens of other baby animals to compare their
performance. In the case of some species, like goats, not a single one would
willingly risk the glass.
And then, in 1985, a team led by James Sorce tried a subtler variation of the
experiment, drawing on the idea that what the parents modeled for the children
might make the difference. They built a visual cliff that presented greater
uncertainty, by being a little less daunting—only roughly thirty inches high. The
researchers used only one-year-old subjects. And this time, they asked the
mothers to act out a variety of emotions in response to the cliff, signaling
approval with big smiles, or changing their faces at a certain point to display
intense fear. When the mothers looked happy, fourteen of nineteen babies
willingly crawled forward over the edge. But when the mothers looked afraid,
none of the babies risked the glass. Most of them backed away from the edge.
Take the powerful, almost narcotic effects that researchers like Christakis,
Domoff, and Niec are working to quantify and struggle against. Couple those with
the long-term effects that watching a parent absorbed in a device may have on
young children trying to sort out their own relationship to a screen. Consider how
powerful childhood memories are in determining what that child will consider
appropriate and inappropriate behavior as a parent later in life. And think about
how long it will take before the research will be there to establish any causation,
much less any parenting strategies or federal safety regulations, even as
generations of children are changed. As adults, we barely register the creeping
effects of technology like surveillance and facial recognition, and how they might
change our expectations about privacy and public life. Meanwhile, children are
growing up in that world, exposed to all these effects from the day they’re born,
without any guidance from their parents as to how it might be shaping us all. This
is The Loop gathering momentum, whirling powerfully just outside our
immediate ability to sense it. And it’s beginning to gather up not just our privacy,
not just our family time, but everything else, too.
Chapter 9
THE LOOP
BY NOW I hope it’s clear that coParenter is part of the final loop I set out to
explore in this book. It belongs to a new wave of businesses deploying endless
products that claim to not just spot patterns in human behavior, but also shape
them. Those businesses are growing rapidly.
Boston-based meQuilibrium uses AI to analyze the digital communications of
each employee in a company to form “a baseline resilience score, profile and
tailored plan to address their areas of greatest vulnerability,” according to the
company’s online marketing materials. Basically, it reads their communications
and makes automated suggestions about how they can improve as individuals and
how the company can better manage all its employees overall.
Cogito, another such company, is a sort of real-time coach for call-center
agents. Its software analyzes phone conversations with customers and looks for
signs of frustration, hurried speech, even “intent to buy.” As the call-center agent
progresses, he might be given an “empathy cue” if the software thinks the
conversation needs a bit of warmth. According to the company’s sales materials,
not only can agents use the technology to “speak more confidently, concisely, and
compassionately,” but the technology also looks for the common characteristics
of the highest-performing agents, and “provides automated guidance to help all
agents perform like the best.” And by looking for patterns like mimicry,
consistency, turn-taking, “a live dashboard shows the behaviors at which agents
are excelling, or highlights those which require attention.” Cogito’s cofounder and
CEO, Josh Feast, says the results are so promising he’s interested in expanding
the technology to other social areas, like communication between coworkers.
(Writing in Fast Company, UNC Chapel Hill professor of law Jeffrey Hirsch
points out that “except for some narrow exceptions—like in bathrooms and other
specific areas where workers can expect relative privacy—private-sector
employees have virtually no way, nor any legal right, to opt out of this sort of
monitoring. They may not even be informed that it is occurring.”9)
And then there’s Affectiva. The company’s CEO, Rana el Kaliouby, first put
AI to work in an attempt to help communication between people with autism and
the rest of the world. But soon she began applying the technology to a broader
and more commercial set of purposes, like scanning the driver of a car for signs of
danger. El Kaliouby explained to me that her AI knows, for instance, that when
the eyes are tired but the mouth has stopped yawning, a driver is likely about to
nod off at the wheel. However, the company’s newest foray is into true emotional
measurement, such as attempting to discern whether someone has had a positive
reaction to an advertisement or to a movie’s pivotal scene. We know, thanks to the
work of people like Paul Slovic and Wendy Wood, that we offload huge amounts
of our decision-making to our emotions, that “going with our gut” is one of our
primary behavioral mechanisms, and that we have almost no conscious control
over any of it. Giving AI a shot at measuring our emotions is a vast business
opportunity, and it paves the way toward influencing our emotions and the
decisions we make with them.
Many critics have argued that these technologies simply can’t do what their
creators claim. UCSF professor of psychiatry Wendy Mendes, who has made a
career studying the ways our emotions shape our behavior, says that the cultural
and individual differences between people when it comes to how our faces
convey anger, or humor, or sadness, make it nearly impossible, not to mention
dangerous, to try to read them through a trained algorithm.
“Am I angry, or do I want you to think I’m angry?” she asked me during a
lunch together in San Francisco. Microexpressions, she told me—the tiny
movements of the face that AI can be trained on—are just too subtle and too
varied for some sort of standardized, automated analysis. “This belief that
microexpressions are a read-out of an inner state is a problem. They’re for
communicating an inner state, and they’re culturally specific,” she says. “Taking
this idea that there’s this perfect, systematic, statistical regularity simply doesn’t
hold up.”
And beyond that, she believes that the promised convenience of a system that
claims to be able to detect intent—the intent to steal or to swing a fist, for
instance—will be too tempting for a police officer or federal investigator. “Law
enforcement agencies will make reverse inferences if this stuff becomes widely
available,” she tells me, because the possible time-saving potential is irresistible,
even though as scientists “we beat each other up all the time in a paper when we
try to make this reverse inference.”
She points out, for instance, the SPOT program, a nearly billion-dollar effort
by the Transportation Safety Administration based on the work of the American
psychologist Paul Ekman. Ekman, who was a pioneer in the study of
microexpressions, helped TSA create a system that relied on a taxonomy of facial
expressions. “When there’s a lot to lose, such as death or imprisonment, the
emotions generated are intense and very hard to conceal,” Ekman wrote in a 2015
blog post defending the program, “so they often leak out.”
“I actually have the endowed chair in his name,” Mendes told me, “and I have
huge respect for him. But taking his work and assuming that the basic emotions
are invariant across culture? That just hasn’t held up.”
In November 2013, the Government Accountability Office released a report
that found that after spending nearly $900 million on the program, the TSA had
no evidence it worked whatsoever. That day, then South Carolina representative
Mark Sanford grilled the head of the TSA at the time, John S. Pistole, rattling
through the potential reasons the SPOT program might pull someone out of line.
SANFORD: If you were a staunch right-wing conspiracist with very strong anti-
government leanings, you’ve posted some things that probably weren’t the
best to post on the Internet, but you had the invisibility that goes with the
Internet, but now you’ve got a law enforcement officer probing, asking you
questions, would you exhibit stress or fear?
PISTOLE: It depends on the individual, but potentially, sure.
SANFORD: You’re an immigrant whose dad and mom came here illegally,
would you exhibit stress or fear if someone was asking you questions?
PISTOLE: All situational, again.
SANFORD: Let’s say you’re a wife whose husband had beaten her and you’re
just trying to get on an airplane and get out of town, would you exhibit stress or
fear if someone was going into interrogation on some front?
PISTOLE: Again, situational.
SANFORD: Which I think raises the point which the GAO report has brought.
You go through a screening system which essentially undresses somebody, you
send their equipment through radar detection and other devices. The question
is, from a civil liberties standpoint, given those other tests, do you in addition
have to go through a screening process based on somebody’s interpretation of
what might be in your brain?
PISTOLE: You raise good points, congressman. There’s no perfect science,
there’s no perfect art of this.
MISSION CRITICAL
ONE MORNING IN 2016, I attended roll call at the Newton Division of the Los
Angeles Police Department. Newton comprises roughly nine square miles and
150,000 people across four neighborhoods (including a public housing project)
just south of downtown Los Angeles. Police work shifts across night and day,
meaning there’s none of the sleepy early-morning vibe of a workplace coming to
life, and even at dawn the halls were full of bluster and activity as one patrol
handed their shotguns and radios across the desk, talking loudly about their plans
after work, and another waited to collect equipment for the start of their day.
Roll call began with a handful of routine announcements, and then the watch
commander listed off a handful of crimes—vehicle thefts, burglaries, assaults—
that a system called PredPol had predicted would take place in their jurisdiction
that day.
The watch commander finished by asking the officers in the room to “spend
time in the PredPol areas,” pointing to a screen behind him, “from the 60s and 70s
up and down Broadway.” Seated at the back of the room, wearing a Beretta and a
tactical vest, Officer Marcela Garcia glanced through the printed map in front of
her, which recommended specific times to roll through specific areas. Then she
checked out a shotgun and a radio at an equipment desk and walked me out to her
Ford black and white, and we settled in together for a drive around the
neighborhood.
PredPol is a for-profit company founded by Jeff Brantingham, an
archaeological anthropology professor at UCLA who created a system with
Department of Defense funding for predicting battlefield casualties on the ground
in Iraq. He adapted those algorithms to make similar predictions about crime,
using COMPSTAT data from LAPD in collaboration with Deputy Chief Sean
Malinowski.
COMPSTAT is a crime-data system created by the NYPD in 1995, since
replicated all over the world. It was once praised for standardizing certain forms
of data for refining police procedures, but in an anonymous survey of more than
1,700 retired police officials, criminologists found rampant manipulation of the
data. “Results indicate that the misuse of the performance management system
and pressures on officers from management are key explanations for manipulating
crime reports.”1 In one 2010 report, an NYPD whistleblower charged that officers
were downgrading felonies to misdemeanors, and in some cases actively
discouraging victims from reporting their crimes, to make their COMPSTAT data
look better. It was one of the first large-scale examples of The Loop, in my view,
when the process of analyzing professional behavior mutated the behavior being
analyzed, and not for the better.
In Los Angeles, Brantingham and a research team went on to refine their
forecasting process, patented what they built, and founded PredPol, LLC. In 2011,
LAPD became one of the company’s first signed clients.
More than 450 UCLA faculty and students signed a letter in October 2019 that
argued
The letter also raised concerns about the possibility that PredPol’s “use of
historical crime data naturalizes policies and practices that have had disparate
impacts on Black and Brown communities.”
By some metrics, PredPol seemed to work. The LAPD Foothills Division
reported a 13 percent drop in crime in the first four months after rolling out
PredPol. The Alhambra Police Department reported a 32 percent reduction in
burglaries and a 20 percent drop in vehicle thefts after it rolled out PredPol. More
than sixty departments nationwide adopted the technology, including big cities
like Seattle and Atlanta. Atlanta, which adopted it in 2013, reported an aggregate
drop of more than 8.5 percent in total crime in the two neighborhoods where it
was deployed.
I sat with Professor Brantingham and a camera crew for an interview and
asked him whether the system he’d built might fall too heavily on innocent people
who happen to live where the math predicts crime will occur.
“There’s a big difference between trying to predict who is going to commit a
crime, and where and when a crime is most likely to occur,” he told me.
I pointed out that a pattern of where and when can very easily lead to the
effect of pointing police at individuals. “I think about a teenage kid who happens
to be living in one of the three houses your system has identified,” I said to him.
“He’s going to be more likely to come under the eye of a police officer or, worse,
get put into a squad car, even if he hasn’t done anything wrong, just because the
system has identified that place as a likely place for crime.”
He disagreed. “The only data that is being used here is what type of crime is it,
where did it occur, and when did it occur. So it’s not focused on the individual,
it’s focused on the events themselves.”
“Although it does zoom in on them.”
“Not on people, on places.”
“On where they live.”
“That’s right—on where they live.”
I spent the morning riding along with Officer Garcia. We settled into the
deeply worn black seats of her squad car, and she drove me through one area, a
half-block at the center of a five-hundred-foot by five-hundred-foot square on a
map, that the software had identified as being likely to see crime at that time on
that day. Officer Garcia told me that PredPol more or less tells her when to relax
and when to be on alert.
“If we see suspicious activity in the sense of a traffic violation in the area, or a
pedestrian violation, we can always do a consensual stop,” she says. “Stop and
speak to them, get information, of who they are, if they live in the area, and what
they’re doing in the area.”
“Do you guys find that you make more arrests within those boxes? Is that one
of the outcomes here?” I asked her.
“That is one of the outcomes,” she says. “We saw a reduction in crime just
based on the amount of time we spent in those boxes.”
I asked Sergeant Garcia effectively the same question I posed to Brantingham:
“Does it ever feel to you like for someone who isn’t breaking the law but happens
to live on this block that we’re cruising right now, that they’re unfairly getting
your eye on them? You’re ready for trouble on this block, they happen to live
here, but they’re doing nothing wrong?”
“We don’t want people to feel that way, at least that’s not our hope for what
they’re gaining out of us. We want them to actually feel good that we’re more
visible in the area in which they live.” It may be that some residents of that
particular block will feel safer by having police come through more often. But for
many, especially after the deaths of Oscar Grant, George Floyd, Breonna Taylor,
and too many others, that’s unlikely to be the case. The objective function doesn’t
benefit everyone.
Professor Brantingham says the point of PredPol isn’t more arrests, just
disrupting what the algorithm predicts will happen by exhibiting police presence.
“It’s been known for quite a long time that there’s something called residual
deterrence, a police officer shows up in a location, and when they leave, their
effect persists for quite a long time.” The scholarship for the most part confirms
this—the threat of longer prison sentences seems to have no effect on crime, but
the threat of apprehension, in the form of more police and more visibility for
them, does seem to have a deterrent effect.2
But what are the long-term effects on policing itself? That part we haven’t
studied yet. Like coParenter, the system is doing the training, and as we’ve seen
with cases like David Dao, we’re not good at saying no to the system. Imagine a
police officer who comes up through the ranks using a system like this. What
effect will it have on her patrol habits, her instincts, the broad discretion most
departments give their officers, if she’s looking to a piece of software to tell her
when to turn her vigilance on and off?
After PredPol pioneered the field, companies like Palantir, HunchLabs, and
IBM followed with their own predictive-policing software packages, all of which
claim, like Brantingham did, that race is not taken into account. But NYU law
professor Barry Friedman, writing in the New York Times in 2018, put it as
follows:
Police may “go where the crime is,” but because so much focus has been
on low-level offenses in disadvantaged areas that are ignored elsewhere,
these algorithms make it inevitable that the police will return to these
places time and again.3
WEAK PERFECTION
HOW DO WE put some boundaries around the use of AI in the domains where
its use really can determine the course of our lives? First, we must recognize the
key assumption we generally make about the wildly transformative potential of
artificial intelligence, because it’s a dangerous one: that it can improve anything
on which we sic it. This has been the moral argument for putting AI into all our
lives. The ability to pick patterns out of a vast jumble of data is, of course, hugely
powerful. And the resulting predictive analysis allows companies to offer the
possibility of getting in front of human behavior, of winning investments, of
possible threats. But we haven’t yet sorted through the possibility that while
robots can turn faster in a fighter jet than we can, that doesn’t mean we should let
them pull the trigger. Similarly, in other critical areas of life, AI might work
against our values, making life more efficient in places where inefficiency is, in
fact, a hidden human safeguard.
California Supreme Court justice Mariano-Florentino Cuéllar, writing in an
extremely prescient 2016 article, pointed out that if we’re going to establish goals
for machine learning’s role in administrating our lives, “those goals must also
inform whether we assign value to the characteristics of human cognition that
contrast with what a computer program could accomplish.”1 Computers are good
at optimizing a system, but it takes humans to sort out what we actually want
from that system.
For example, the system of law could of course be made vastly more efficient.
“Think about entering a plea,” Cuéllar, now head of the Carnegie Endowment for
International Peace, told me. “It’s a choice you only get to make once—there’s no
taking it back—and that plea has the potential to change your life forever.”
Entering a plea in California is a rigidly regulated process. The California penal
code stipulates that pleas must be entered in open court, for one thing—no doing
it remotely. And if I’m going to plead guilty or no contest, the court has to
provide me additional time to consider the implications. And if the offense carries
the risk of deportation, a life sentence, or other extraordinary punishments, all
sorts of warnings and negotiation processes and approvals from my lawyer
become necessary as well.2 Pleading guilty or not guilty is extremely
complicated.
Could technology make it simpler? Of course. We might reduce the whole
process to a single swipe on a smartphone, a world in which defendants
conveniently and casually enter their plea one-handed from a cell, or while
waiting at a stoplight. But that would be disastrous, according to Cuéllar. “The
legal process puts people through an intentionally slow decision-making system
when they decide on their plea. It forces slow thinking. We call it ‘weak
perfection,’ the idea that the system remains difficult to help them make better
decisions.”
When I think about how I’d want pattern-recognition algorithms to be
regulated, I think about how “weak perfection” describes a whole host of systems
that are slow, inefficient, fragile, difficult to navigate—and force us to stop and do
the most mentally taxing kind of System 2 processing. Perhaps sometimes we
should have to do it the hard way. We should have to write out our specific desire
not to be intubated at the hospital if we fall into a coma. We should have to sit and
decide how we want our money to be distributed among our children when we
die.
Meanwhile, AI is making its way into certain systems of judgment to make
them far more convenient for the bureaucracies they serve, and in exchange we’re
losing our insight into how important decisions are being made. AI can find
patterns and form predictive rules, but as we’ve seen, the decision-making
mechanism is hidden from view. Professor Michele Gilman, who teaches law at
the University of Baltimore, has spent two decades suing on behalf of people
living in poverty who have been denied their rights and benefits. And
increasingly, she’s found that algorithms are at the center of her cases. In a 2020
paper for Data & Society, she offered a guide to the algorithms now being
deployed to make decisions in literally dozens of areas of American life, from
credit scores to public benefits to housing. Gilman points out that algorithms
often make life-changing decisions without directly consulting the people whose
lives they’re changing. “An applicant for rental housing, for instance, may be
denied based on an algorithmic determination in a tenant screening report,” she
writes, “yet never learn about the basis of the denial.”3
Gilman identifies several increasingly alarming layers of opacity. Except in a
few progressive jurisdictions, like Seattle and New York City, which require
transparency and public consultation before an algorithm can be adopted,
government agencies usually activate these decision-making systems with little or
no public input. And she points out that although due process usually requires that
the companies making algorithmic products have to explain them in court, that
tends to involve a fight because those companies usually force government clients
to sign nondisclosure agreements and claim trade-secret protection when asked
about the inner workings of their AI.
In fact, just standing in court, charged with a crime, you or I would likely be
under the gaze of an algorithm. Pretrial risk assessment tools (RATs) are in use in
forty-six states and the District of Columbia, according to a report by the
nonprofits Media Mobilizing Project and MediaJustice. These tools were
introduced as a means of fighting off the inherent inequality of cash bail, which
disproportionately keeps poor Americans in jail, and as a means of compensating
for the possible biases in a judge’s discretion. (One influential National Bureau of
Economic Research study found not only that Black Americans were
disproportionately stopped, searched, arrested, and sentenced harshly, but that
those racial biases were consistent in both Black and white judges.4)
RATs look across variables in a population to offer a statistical prediction of
the likelihood that a defendant is a flight risk or might commit another crime
while free awaiting trial. Those variables might include demographic information
like age, access to housing, and job history, and might also include criminal
history, such as substance abuse, prior failures to appear in court, and time spent
in prison. The software produces a risk score for each defendant so that judges
have a handy way of evaluating whether to release them pending trial, ostensibly
free of their own biases.
But as several investigations and research projects have found, RATs can be
deeply biased themselves, whether because they channel the mathematical
inequalities inherent in American life into their judgments or because they create
an overly broad judgment from millions of cases that may not apply to a local
court and the conditions facing defendants there. Even studies that have found
minimal racial biases in RATs come with stark warnings about overreliance on the
software. “Risk assessments are believed to remove conscious and unconscious
forms of human bias, and provide a system to treat people fairly,” wrote the
authors of one Criminology & Public Policy study that found RATs don’t
themselves wind up having an unfair impact on defendants. “However, there is
nothing inherent in risk assessments that will reduce jail populations, make prison
populations less racially disparate, or otherwise reform the criminal justice
system.”5
Some who hoped RATs might unwind bias in the criminal justice system have
now come to actively argue they shouldn’t be used at all. Back in 2014, when the
state of New Jersey was considering whether to adopt RATs, an endorsement
from the nonprofit Pretrial Justice Institute helped to persuade lawmakers to use
them. But in 2020, PJI, whose mission is to improve the equity of the criminal
justice system in how it evaluates defendants, wrote in a statement that it had
been wrong. “We heard but did not fully appreciate the opposition to pretrial risk
assessment tools from civil rights organizations, impacted people, and
researchers,” the statement read. “Regardless of their science, brand, or age, these
tools are derived from data reflecting structural racism and institutional inequity
that impact our court and law enforcement policies and practices. Use of that data
then deepens the inequity.”6
The tendency to broadly apply the rulings of artificial intelligence when
human systems are overwhelmed by human need damages independent judgment
everywhere—not just in the courts. And it’s when we strap AI and high
expectations onto a broken system that things really go wrong.
By January 2021, ten months into the COVID-19 pandemic, the economic
fallout in the United States was severe and deepening. The service industry was
decimated as restaurants, hair salons, and gyms faltered and closed. The head of
the National Association of Consumer Bankruptcy Attorneys told me his
members were preparing for an avalanche of lost homes and cars as American
entrepreneurs everywhere threw whatever they could at saving their businesses.
And as the New Year began, with more than 10 million jobs gone from the
economy, more than 9 million Americans fell into poverty.
At that point, a stimulus bill was supposed to begin sending checks to help
Americans in need. But because the administration of that relief had been handed
off to the states by the Trump administration, state unemployment offices were in
charge of distributing those funds. In California, that agency is the Employment
Development Department, and like all such agencies across the country, it was
utterly overwhelmed.
In 2010, the worst year of the prior recession, EDD saw 3.8 million
unemployment claims. But the coronavirus brought more than 16 million claims,
and as of October, 600,000 calls per month meant waiting on hold for an entire
working day in the hope of getting through. And then, in November, news broke
that EDD had been accidentally sending benefit checks to people in prison. It
turned out that since the beginning of the pandemic, nearly 40,000 inmates had
filed claims, federal and state prosecutors found, including 133 death-row
inmates, and more than 20,000 such claims were paid. All told, the state had
wrongly given out more than $140 million to incarcerated people. (Some might
argue that incarcerated people need additional resources, but surely distributing
millions by mistake only to those who thought to illegally apply for them isn’t the
way to do it.)
The EDD, reeling and embarrassed, and without the political support to ask for
enough humans to clear the backlog and process new claims, decided to use a
risk-detection system.
The first sign that this new system was going to pose problems for everyone
was a tweet by EDD on a Sunday in January 2021: “As part of ongoing efforts to
fight fraud, EDD has suspended payment on claims considered high risk and is
informing those affected that their identity will need to be verified starting this
week before payments can resume. More details on the EDD website in the days
ahead.”7
I asked EDD how they were determining high-risk claims. They told me they
did it by throwing an off-the-shelf piece of software at the problem. “Using the
Thomson Reuters fraud criteria, EDD reviewed existing claims and applied
additional industry standard fraud detection criteria, taking action on claims
deemed highly suspect or fraudulent,” they wrote. “In order to mitigate future
fraudulent payments, EDD has stopped payment on those higher risk claims and
has communicated to those affected that we will be requiring identity verification
or additional eligibility determinations starting later this week before payments
can resume.”
How many claims did the Thomson Reuters product flag? One in seven. More
than 1.4 million claims were frozen by the system, meaning those Californians, at
a time of extraordinary hardship, would have to navigate additional bureaucracy
to prove they were eligible to collect a weekly payment of $300 from the federal
government. The process reeks of misplaced assumptions. EDD assumed that
software could better decide on a course of action than could humans. The
software assumed that one in seven Californians were in some way undeserving
of immediate aid. And EDD assumed the software was right. As we hand more
and more crucial decisions to automated systems, the assumption that computers
can do things better than humans can is going to become a matter of life and
death.
Perhaps we’ve simply gotten ahead of ourselves. Perhaps inside The Loop
we’re just too quick to measure the benefits of AI in time and money—and avoid
uncomfortable choices in the process. Perhaps where AI should be applied is not
in issuing predictive judgments, which creates deep inequities that are difficult to
untangle once we’ve built a human system on top of them, the way EDD did.
Instead, the same way we’ve considered using AI as a potential way of looking at
what’s wrong in policing, what if we use AI to assess our ugliest hidden patterns
in all parts of life?
In 2018, I toured the Metropolitan Museum in New York with its chief digital
officer at the time, Loic Tallon. The museum was in the process of digitally
scanning its collection as part of a vast initiative to make it easier to experience
both in person and online. For example, as part of a partnership with Microsoft,
the Met was beginning to make recommendations to visitors as to what they
might enjoy next, the way recommendation algorithms identify the next item
you’re likely to buy on Amazon. Tallon and I wandered through the Byzantine
section and riffed on what AI might be able to do with the collection as a detailed
data set. “We’ve tagged 5,000 years of human history,” he told me proudly. And
while making recommendations to visitors is cool, he said he was excited about
the possibility that the museum could use machine learning to “genuinely study
the collection itself, and start finding trends and pointing to particular things that
the AI may spot across multiple centuries of content.” For instance, he theorized,
the AI might be able to fill in certain evolutionary gaps in the history of art.
Pottery, he pointed out, is so inherently fragile that there are huge periods of time
in which we simply don’t have representative samples. In the time between
Egypt’s fall and the rise of the Greek state, one jug and the next known example
could be more than a century apart. But AI could look at the materials, the
painting, the proportion, and speculate as to what might have come between the
two. “It feels inevitable that at some point the AIs will be sufficiently trained, the
data sets will be sufficiently large and rigorous, that genuinely you’re going to
find those patterns,” Tallon said. We won’t be using AI to sell art, or to make it.
We’ll be using it to better understand our own creative impulses, what
connections our society tends to make between aesthetic and artistic movements.
We’ll uncover the hidden topography of human history.
Whole worlds of scholarship have been waiting for this sort of tool, a way to
rigorously identify patterns in our past as clearly as AI identifies cancer in
millions of MRI scans. But can we somehow give it to the people who can make
ready use of that tool in studying human beings, without also giving it to
companies that want to use those same studies to identify markets and sell
products?
AT THE END of a long, isolated gravel road in the woods of Virginia, there’s a
quiet couple that studies the very worst human behavior. The historian Roberta
Senechal de la Roche is a professor at Washington and Lee University in
Lexington. She’s a poet and has written a book about Civil War snipers (an
ancestor was a Union sharpshooter), but her life’s work is the analysis of
collective violence. She specifically studies the group decisions that have led to
lynching, race riots, and terrorism, and the conditions that made lynchings so
common and race riots so unusual in the South, while the opposite was true in the
North.
Her husband is the University of Virginia sociologist Donald Black, who
studies what he calls “the behavior of law”—how it reflects ancient tendencies
and why it falls so heavily on the least-powerful members of society. I have been
fascinated with the two of them for years, because to me they embody two things
that are both contradictory and complementary. On the one hand, they seek to
study long-term patterns in human behavior, because as a society we’ve been
terrible at seeing those patterns clearly. They know better than anyone the myopia
that will keep us from spotting technology’s slow, steady influence on how we
live until it’s too late. On the other hand, their work is the sort of scholarship that
depends on more data than humans can possibly process. If these two were given
pattern-recognition systems, and applied them to the history of lynchings, or the
history of law, histories that involve more massive amounts of data, the work
could transform our understanding of ourselves. But people are using AI to
recommend eyeglass frames and vintages of wine. No one’s hoping to make
billions of dollars by paying to train AI on the history of violence.
Black’s work in particular could benefit immensely. “I’ve always been
interested in the contrast between the tribal way of life and the modern way of
life,” Black told me. “These clashes of ‘right’ and ‘wrong’”—the crimes we have
committed and punishments we have meted out throughout history—“how do
they change between the tribal and the modern world?”
Black is not a modest person. His theories are vast and sweeping. An
academic journal once asked to interview him, and because, as he told me, no one
else was qualified to interrogate his ideas, he submitted a self-interview. And his
work, which he says draws on evidence from all over history and all over the
world in service of theories that apply to all of humankind, was called by one
frustrated critic, the NYU sociologist David Greenberg, “a terribly inefficient way
to generate the elements of a theory.”8
But there is something haunting in what Black has spent his career describing,
and if AI were brought to bear on it, it might be extremely valuable. His theory is
roughly this: the way we assess crime and dole out appropriate punishment today
is a translation of ancient moral codes. Inside those codes are some very
fundamental assumptions and biases when it comes to power, gender, and tribal
identity. And those codes play out over and over again in how societies deem
certain behavior criminal and accord it more or less punishment. The overall rule
of thumb has to do with what he calls “relational distance”—how close or far
from the center of a society’s tribal and moral identity are the perpetrator and the
victim. The severity of the punishment increases as the perpetrator becomes more
distant, and as the victim becomes closer to the center of the society and its power
structure. At first what he describes seems head-smackingly simple. The killing of
animals is punished much less severely than the killing of humans, for instance.
But as he proceeded along that path, he found that the consistency with which
some crimes are punished more or less harshly reveals how societies rank the
value of the people committing crimes, and the people suffering them.
“I don’t know of a single case of a man killing his wife in which the man
received capital punishment,” Black told me. “I mean, there must be some case
somewhere in the history of capital punishment in the United States, but I’ve
never seen one. You have [capital] cases where a man kills, say two or three
members of his family, but not just one, not just a spouse.” The code by which we
punish men who kill their wives tells us something horrible, Black told me, about
how we feel about marriage, and the privileges of men, and what dominion they
get to have over their wives. Black says it may also suggest we prize civil
relationships between strangers, and perhaps the lives of children, over the life of
a married woman. “Killing an intimate attracts drastically less severity than
killing a stranger,” Black said. “Capital punishment is basically a punishment for
stranger killings, plus multiple killings.”
And law, according to Black, has its own gravitational rules. Pour it from a
higher sociological height—from a more highly placed victim in society down
toward the perpetrator—and it falls more heavily. People who commit a crime
against the state are far more severely punished than when the state commits a
crime against citizens. Black men convicted of federal crimes receive 20 percent
longer sentences than white men convicted of the same crime, according to a
report from the United States Sentencing Commission, and that disparity is only
growing.9 And the race of the victim is hugely consequential. A Northeastern
University School of Law review concluded that “regardless of the perpetrator’s
race, those who kill whites are more likely to face capital charges, receive a death
sentence, and die by execution than those who murder blacks.”
When I asked around among law professors and sociologists, Black has the
reputation of being on to something without having produced the evidence
necessary to support it. “It’s the kind of thing where he might be right, but he
can’t prove it,” one sociologist told me.
Black says he’s frustrated in his work, because he feels he’s nailed a really
compelling theory, but not enough people are paying attention, and he doesn’t
have the tools he needs to flesh it out. “People say ‘but why is that? Why does
more relational distance attract more law?’” he said. “I can tell you why the cases
are handled the way they are with this theory. I can’t tell you why the social
geometry of the cases predicts the way the cases will be handled. Maybe someday
someone will figure out how.” He sighs. “I feel like if some lab professors could
just understand it, they could make a whole damn career off of it.”
What Black has inadvertently done is build a theory the same way AI might do
it—he’s taken a pattern-based shortcut to an answer and has built a broad theory
on it. But now he’s facing the explainability problem. He doesn’t have what it
would take to go inside and figure out why his theory is seemingly so correct, or
what else it might predict.
In a court of law, the most powerful use of AI might be as a mirror, one that
shows us the mathematical patterns of inequality we’ve perpetuated across
hundreds of years and millions of cases. And if a hypothesis like “relational
distance” is borne out, perhaps that becomes the basis of a recommendation to a
judge, rather than the sort of systems that now dress the brutal mathematical
inequalities of modern life as impartial, time-saving judgments, and pour them
down onto defendants.
That sort of problem might be just the thing for AI to accomplish. Feed an
algorithm enough of the history of criminal justice in the United States, and it
might very well begin spitting out certain predictive connections. Perhaps a
theory like Black’s can be investigated the way AI might identify the missing
links in the history of art. If we’re going to blindly believe what AI has to say
about something, maybe it should be about who hasn’t received justice, rather
than who should. Perhaps that needs to happen before we put AI to use in
determining whether someone gets out of jail tonight.
Chapter 12
HIGHER MATH
UNTIL NOW THIS book has offered a few theoretical solutions to all the
problems The Loop presents. They’re invisible problems, years out, perfectly
tailored to our specific human biases and heuristics, and to the business incentives
of capitalism. It’s a dark picture, I know—I’ve been ruining dinner parties with it
for years. But I want to end the book with some possible solutions. Here we’ll
look at a few projects that harness the same technology and business dynamics
that make The Loop possible, and could conceivably be the beginning of a Loop
from which we might actually benefit. “Harness” is the term because until now
we’ve allowed these forces to simply run wild. But with some work—a hard look
at how the brain does what it does and at the things it doesn’t do well, as well as a
similar inspection of AI and of capitalism—I think we can build some rules
around this thing and break the circular pattern of guidance-for-profit. It looks
impossible, but there are shoots of green already appearing across the landscape.
After years of record-breaking wildfires across Northern California, major
insurers began to drop their coverage of certain homes outright. By the math of
their risk-assessment systems, the gamble simply wasn’t worth it. A company
called Verisk, which dominates the data-analytics market for insurance
underwriting, estimated in 2020 that 15 percent of properties in California were at
high or very high risk for fire.1
The traditional means of assessing fire risk is a low-resolution method:
average risk across a broad number of homes, whether by evaluating census tracts
or entire zip codes. And by that system, there’s little or no room for negotiation,
because an insurance company can’t afford to send someone out to every home
multiple times to check out what fire-proofing improvements have taken place.
All of this meant that a homeowner who had cleared away brush and thinned trees
and put fire-safe vents around her home was being lumped in with her neighbor
who still had dense, flammable brush everywhere and siding and roofing that
made his home likely to catch when the next embers floated down from a forest
fire nearby. Insurers had no choice but to lump the two together, based on
historical patterns of fire risk, and deny them both.
But Met Life, Farmers, and a few other companies have begun quietly offering
more insurance in Northern California communities that are at risk of wildfire.
How? They’ve bought technology from a company called Zesty.ai, which uses
machine learning to find the specific factors that endanger a home in a wildfire-
prone area.
Attila Toth, Zesty.ai’s CEO, says the current risk-assessment models used by
insurance companies are outdated. “Of the five largest loss years, four are in this
decade, but the technology to understand those sorts of disasters is thirty to forty
years old.”
Toth’s company feeds satellite imagery and property databases through
machine learning and builds a recipe for predicting loss. “We pick up thirty
property-level modifiers, things like roof material,” he says. “We measure
overhanging vegetation from an aerial image. We look at the last building permit
pulled for that roof. Slope, vegetation, precipitation, orientation of the property.”
Then, with that recipe, his AI builds a predictive model that gives Met Life an
automated system for offering specific assessments on a property by property
basis. Remember that UC Berkeley study, which theorized that one of the ways to
fight “performativity” in AI was to stop making all predictions on a backward-
looking basis? Instead, the authors wrote, AI should evaluate its predictions
“against the future outcomes that manifest from acting on the prediction.” That’s
what’s happening with these homeowners.
The result is that insurance, rather than being a binary up-down vote on you
and your house’s chances of disaster, could become a means of spurring humans
to take better precautions. Your trees are too close to the house, your insurer
might say. There’s dense brush inside the property line, the flashing around your
roof isn’t fire-resistant, and so here, here’s your estimate. But if you bring a
chainsaw and a contractor to bear on your fire risk, here’s a vastly more
affordable quote.
Toth says that his AI-driven model estimates only 2.5 percent of properties are
at risk, as compared with Verisk’s estimate of 15 percent, and while a narrow slice
of that group may have to be prodded by lack of insurance into living elsewhere, a
lot of them may be able to take the steps his AI and his inspectors have identified
to lessen the danger. “Even if a property falls within the fire perimeter,” Toth
says, “if you have defensible space, if it’s well-maintained, if it’s oriented in the
right direction, the probability of that house burning is very low.”
Of course, this doesn’t solve the larger problems of risk and analysis and
profitable operations that plague the insurance industry. It’s not as simple as
saying that we need to use more humans in the process, for instance—that’s just
not feasible. And generally speaking, an industry built on using math to aggregate
human behavior into pools of risk, and then betting on the outcomes, is not
looking for exceptions, it’s looking for rules. It doesn’t want to tailor a suit for
each customer; it wants to sell one suit. And so far, none of the companies using
Zesty.ai are issuing tailored suggestions to their clients—you have to know to
push for it when they issue you a quote or turn you down. But in a world in which
the climate is changing, fires are becoming more frequent and more destructive,
and people are doing their damnedest to adapt, an insurance company that only
uses historical data and lumps everyone into the largest possible pools is signing
up for risks no balance sheet can handle.
It’s not at all clear that whatever companies do with something like Zesty.ai’s
system will make life easier for everyone. The arc of history doesn’t bend in a
satisfactory direction for all customers. Many of the people who live in the path
of wildfires don’t have the resources to bring in a timber team to trim back their
trees. Further, an automated, tailored set of “or else” recommendations may force
them to either live without insurance and risk losing everything when fire comes
through, or give up the house ahead of time. And how do you sell a home that
cutting-edge technology has determined is currently indefensible?
But this is a case where some forward-looking, AI-powered analysis can help
us perhaps do what we’ve never been able to do before: encourage people to think
about their risks before they buy or build a home. There is currently so little
regulation as to where one can build a rural or semirural home in the United
States, and there is so much climate-fueled catastrophe coming toward us,
perhaps an AI-driven system for determining risk is a necessary piece of
guidance. In California, homes are regularly built in the path of recurring
wildfires, in the shadow of levees across the Sacramento–San Joaquin Delta, or in
low-lying oceanfront communities in Los Angeles, all of which are subject to the
intensifying rhythm of fire and flood and coastal storms. Zesty.ai could apply its
technology to any disaster-prone location in the world. And if it did, perhaps we’d
wind up pulling people away from places that simply aren’t safe to live any
longer. AI can help sell a home by matching historical patterns, reading emotions,
even writing advertising copy, sure. Those are the forces of The Loop. But if AI
can help dissuade us from buying home in a dangerous place, or ensure that we
don’t sign a mortgage on a home without understanding that we’ll need to spend
another $10,000 fireproofing it, that’s something else.
THE LOOP IS going to fool us into believing that AI is worth adopting for the
time and labor it saves, and the case for its use will be made again and again in
the numbers. A Facebook CIO famously used the phrase “data wins the
argument” to describe the deliberative process inside the company. But I believe
that we should, as a society, prioritize certain things above expediency, and that
AI cannot be trusted to absorb our values in any reliable way, because often
acting on our values is going to require acting in inefficient and costly ways.
Frankly, doing the right thing often feels wrong, which is why there are other
ways (courts, regulations, philosophy) to evaluate the decisions we make as a
society, and before we go handing much more of our lives to AI, we should make
sure we’ve taken stock of who we wanted to be in the first place.
Sometimes we can express our values as math. In general, America has relied
on simple mathematics to determine the value of a human life, for instance.
Federal agencies in fact price human lives at a fixed rate—the VSL, or value of a
statistical life—as they calculate the financial impact of a regulation. (EPA’s rate
is $9.1 million per life, whereas the Department of Transportation pegs it at closer
to $9.6 million. Strangely, VSL in other countries is lower, presumably because
Americans are more litigious, or perhaps because we enjoy such an enormous
GDP. A New Zealander killed in a car accident is worth less than half what an
American is according to the NZ Transportation Authority.) There are cases
where the gravity of our loss is expressed clearly in the math, and the remedy is
easy to measure.
But then there are instances where the gravity of loss isn’t just a matter of raw
numbers, it’s a matter of secondary losses, or even of emotional impact. As my
friend and producer Carl Byker pointed out while we were filming a documentary
together, we have decided, as a society, that certain extraordinary remedies are
necessary when a loss is just too much—emotionally and psychologically
speaking—for us to bear.
Sometimes it’s a matter of building unique exceptions to our rules so that our
civilization can continue to enjoy the benefits of a thing that in rare instances
hurts individuals. Vaccines are an example.
I want to be clear, up front, that I think vaccines are the greatest lifesaving
invention in human history. This isn’t just because they’re saving lives from
COVID-19 all over the world as I write this. Whether it’s an inoculation against
polio or tetanus, only water purification and sewage systems have served to fight
off death and debilitation as well as vaccines have. Life without them would be
short and end horribly for billions of people, and being vaccinated is the ultimate
act of communal compassion because it protects not just your body, but the bodies
of your family, neighbors, colleagues. When a vaccine passes clinical trials and
makes it to market, get the shot.
That said, there are rare cases in which people are harmed by a routine
vaccine, typically because of a severe allergic reaction. How rare is this? The
CDC estimates that for most vaccines the rate is literally one in a million, and in
most cases the reaction is anaphylaxis, the swelling and breathing obstruction
allergic people get from a bee sting. That usually happens within four hours, and
that’s why the healthcare providers who administer vaccines are equipped with
ready shots of epinephrine, which can reverse the allergic reaction almost
instantaneously.
But there are cases in which something like thrombocytopenic purpura (a
crash of one’s blood-platelet count, leading to bruising or uncontrolled bleeding)
or Guillain-Barré syndrome (a debilitating form of nerve paralysis) results from
vaccines. And this puts our society in a bind. The number of deaths and injuries
from routine vaccines is exponentially fewer than the deaths and injuries from the
diseases these vaccines prevent in the rest of the population. But the injury to this
extremely small number of people each year is bad enough (and, as it typically
happens to children, emotionally horrible enough) that conventional liability law
would make it impossible for any pharmaceutical company to continue taking the
risk of producing these vaccines. Still, we need those vaccines. The billions of
lives saved each year means we have to keep them in production. And so we’ve
set up a separate legal system entirely, a carved-out exception to run-of-the-mill
capitalism, in order to make vaccine production legally viable.
If, as a parent, your child is injured by a vaccine, and you choose to pursue
damages, you can apply for remedy at the National Vaccine Injury Compensation
Program, which hears cases at the Howard T. Markey National Courts Building,
on a leafy street in Washington, DC, that looks out over the grounds of the White
House. There, a “special master” will preside over your case, hear your evidence,
and, assuming that medical staff from the Department of Health and Human
Services find that you fall into the proper category, award you a payment, almost
immediately, for your and your child’s loss. The payments cover medical and
legal expenses, future wages lost, up to a quarter-million dollars for pain and
suffering, and another quarter-million, God forbid, in cases of death. Even those
who file an unsuccessful claim usually get their legal expenses covered. It’s an
extremely efficient and sweeping system. The program has paid out more than $4
billion since it began in 1988. And it’s a radical departure from the typical rules of
American business and tort law. The vaccine makers (and American citizens)
fund the pool of money through a 75-cent tax on every vaccine dose administered.
And in exchange, those manufacturers enjoy a “no fault” status in those cases,
avoiding legal liability for making vaccines that have harmed someone, so
everyone else can live in a vaccinated society.
Not only is all of this a radical departure from how capitalism handles nearly
everything else, it reflects a very new way of thinking about what children mean
to us. The sociologist Viviana Zelizer notes in her book Pricing the Priceless
Child that the death of a child was considered until the 1930s a comparatively
minor hardship. Today, although a child typically doesn’t work a job the way
children had to back then, a child’s life is considered vastly more valuable,
against all economic logic, and compensation for the loss of a child today reflects
that illogical yet inarguable society-wide consensus. The point is, we can invent
new systems for new values when it’s important enough to do so.
These compensation programs that do an end run around capitalism are a
model for the sort of thing we’ll have to invent for new forms of technology. Back
in 2007, when self-driving cars were first becoming technologically viable, I
interviewed legal experts on the implications of a vehicle that would so clearly
shift legal responsibility for any accidents from the human driver to the
manufacturer. Even then, the experts I spoke with were convinced that robots
would inevitably be better at driving cars than humans. (One World Health
Organization study estimated that drivers are at fault in more than 90 percent of
accidents.)
“If,” as a car manufacturer, “you can drop the number of accidents by taking
over control, you’re going to face fewer lawsuits,” Ryan Calo, who was studying
the issue at Stanford at the time, told me. He said he thought car makers would be
happy to take legal liability because the pool of liability would still be smaller
than it is today. “If you reduce accidents by 50 percent, who cares if you hold the
bag on the remaining lawsuits? You’ve still saved yourself a lot of money.”
If we get the technology right, and the number of accidents would be so much
fewer, perhaps it’s worth thinking of autonomous vehicles as an inoculation
against human error, a literal vaccine against bad driving. We have a vaccine
court for flu shots. Perhaps we’ll have to build a vaccine court for robot cars and
other forms of AI-based technology.
There are instances, however, when human beings have decided that in spite
of the math, we have to build new systems for society simply because living
without them is just too horrible, too emotionally unacceptable, to consider.
In October 2002, a father and physician named Scott Gulbransen did what I do
each evening when I come home—he backed into his driveway to ensure he’d
have the best possible view of neighborhood children when he pulled out into the
street the next morning. But he didn’t see his son Cameron behind the car as he
pulled in, and killed him. I’m writing this in the most clinical language I can,
because frankly, it touches my worst fears as a parent, and I can’t bear to go into
any more detail than that. Evidently Gulbransen’s story touched lawmakers’ fears
as well, because in 2007 Congress passed the Cameron Gulbransen Kids
Transportation Safety Act, which obligated federal highway officials to figure out
a rule for the design of cars that would better protect kids like Cameron.
The reasoning may have had to do with the growing realization that anyone
can make the same horrific mistake Scott Gulbransen did. In a 2014 test by the
Insurance Institute for Highway Safety, 111 volunteers were given the keys to a
new Chevy Equinox and told to play with the infotainment system while parking
the car. They were then told to back out and drive to their own car. Before they
did, unbeknownst to them, a foam dummy the height of a small child was placed
behind the rear bumper as if a toddler had wandered into the driveway. Without
realizing it until it was too late, every single one of those 111 volunteers backed
over the dummy.
It took more than a decade and a lawsuit by consumer advocate groups to
finally get a rule in place, but as of 2018, any new vehicle sold in the United
States weighing less than 10,000 pounds must, by law, have a backup camera
installed. We recognized a terrible thing, determined that humans are inherently
bad at avoiding it, and we decided not to just blame it on individuals. We
recognized a society-wide problem. And perhaps most importantly, we decided
that the data did not win the argument. If anything, the math actually obscured the
problem.
Because here’s the thing: the number of deaths prevented by backup cameras is
comparatively small. A 2010 study by the National Highway Traffic Safety
Administration found that backover accidents were at that point causing 292
deaths and 18,000 injuries each year on average, and that the new technology
would prevent an estimated 95 of those deaths and 7,000 of those injuries. In
2018, the year backup cameras became mandatory, more than 17 million cars and
light trucks were sold in the United States. Each of them carried a camera, and
each of their buyers paid a little extra, in order to prevent fewer than 100 deaths.
As a society we decided that 100 children being run over by their own parents is
unacceptable, and that we should take steps to reduce that number, small as it is.
Our emotions—our feelings as parents, the value we place on young life, the
realization that any of us might run down our kids—won the argument.
On the other side of the coin, emotions sometimes override the data in other
ways. The same year that backup cameras became mandatory, nearly 40,000
people lost their lives to a firearm in the United States, the majority—61 percent
—by suicide. There was a record 323 mass shootings, including the Parkland,
Florida, shooting at a high school that killed 17 students. No substantial federal
legislation that year or since has changed the way the average firearm is
constructed, or how their sale is regulated.
But sometimes we do it right. Who would not be moved by the horror that
Cameron’s parents experienced, or the simplicity of the problem they asked
Congress to solve? Scott Gulbransen later wrote, “Cameron had died a sudden
and horrible death because he was too small for me to see him behind my
vehicle.” What I’m trying to articulate here is that statistics—the data that AI
would use to decide what’s important—aren’t always a reflection of who we are
or want to be. We shouldn’t always go to the top of the volcano because a guide is
ready to take us there. We shouldn’t plow ahead and deploy profitable technology
on one another that we know has a widespread effect when we don’t yet know
what it is. We are capable as a society of making choices and rules based on
something softer, messier, more important than the data, something we can all
sense but can’t quite see, something that touches on the human emotions we all
share, the immeasurable but unmistakable dark matter of our minds.
NEW YORK HARBOR, where the West Village hits the Hudson River, was
once a chaos of barges. (The main thoroughfare that ran along it, 10th Avenue,
was known as “Death Avenue” for the New York Central Railroad cars that would
roar back and forth across pedestrian crossings, and in spite of the men hired to
wave red flags from horseback to warn of oncoming trains, hundreds of New
Yorkers had died under their wheels by the early 1920s.) And on a cold January
morning in 1944, the tug Carroll was working to retrieve one of a row of barges
moored to Pier 52. Barge work has always been dangerous, but at that time it
carried the extra risk of being highly improvisational. When the Carroll
approached Pier 52, her master found the one open parking space blocked by a
line connecting the pier to a row of six barges, and he sent a harbormaster and a
deckhand to clear the way.
The two men checked the fastenings in the cold, and then threw off the line
blocking the Carroll’s way. Suddenly, all six barges, each more than 190 feet
long, were no longer connected to the pier and drifted south. The Carroll and
another tug tried to rescue them, but one barge, the Anna C, was pierced by the
propeller of a nearby tanker and sank with a full load of flour owned by the US
government. At that time, three months before D-Day, the war effort couldn’t
spare 1,500 tons of flour, much less a barge. So the government sued, and three
years later, in January 1947, everyone involved was in front of the Second Circuit
Court of Appeals, awaiting a ruling by Justice Learned Hand.
Justice Hand had struggled, he said, to determine whether the owners of the
various barges, including the Anna C, should be responsible for paying someone
to remain on board, so they could take charge if the barge broke away. He
considered a case from a century prior in which an unmanned schooner broke
loose and sank another ship. He looked at the exoneration of a powerboat owner
for having no watchman aboard when a group of boys untied it for kicks in the
waters off Brooklyn and it struck another vessel. He was trying to figure out
whether a company that inadvertently unleashes something into shared waters
should be required to keep better control of it ahead of time. “There is no general
rule to determine when the absence of a bargee or other attendant will make the
owner of the barge liable for injuries to other vessels if she breaks away from her
moorings,” he wrote. And so he came up with his own. Today, it’s known as the
Hand Rule.
He described it this way.
Since there are occasions when every vessel will break from her moorings,
and since, if she does, she becomes a menace to those about her; the
owner’s duty, as in other similar situations, to provide against resulting
injuries is a function of three variables: (1) The probability that she will
break away; (2) the gravity of the resulting injury, if she does; (3) the
burden of adequate precautions. Possibly it serves to bring this notion into
relief to state it in algebraic terms: if the probability be called P; the injury,
L; and the burden, B; liability depends upon whether B is less than L
multiplied by P: i. e., whether B > PL.3
“A lot of courts just didn’t get it,” Edelson says. Sure, people were spending
more than they could afford. But in a free market, that’s not considered a loss.
That’s just financial carelessness. That’s the customer’s problem, not society’s.
Telling stories like Wilkinson’s, however, has helped Edelson make the case that
something new is at work here, a pattern-recognition system that seeks out people
whose demographics and online behavior suggest they’ll be particularly
susceptible to addictive play, and, once they’re on the hook, does everything it
can to keep them there. He points to the tactic, used by several social casino
companies, of assigning “VIP hosts” to manage the relationship with the player
and keep them coming back. “You have individual clients who’ve lost hundreds
of thousands of dollars and have said to these casinos, ‘I’ve got an addiction
problem, please stop.’ And then they have a specific concierge who sends them
flowers on their birthday, who say, ‘What can we do to get you back in the
game?’ Those are the stories that we’re going to tell before the jury. And we’re
very confident it’s going to resonate. Once we get these cases to the jury, we think
that the casinos are in a lot of trouble.”
True to his point, Edelson has begun to win. In a case against Big Fish Casino,
his clients received a settlement of $155 million. With Wilkinson and her fellow
plaintiffs, he says he believes he’ll win again. And he says he has a new suit in
the works against Apple, Google, and Facebook for essentially serving likely
addicts up to social casino makers.
Why? Because those platforms are how game makers find their best
customers. The most successful social casino game makers advertise on Facebook
and offer a product that’s playable on any smartphone. Eleven of the twelve top-
grossing games on Facebook, according to Facebook’s own data, are casino
games.6 Wilkinson first encountered DoubleDown Casino, Facebook’s third-
highest-grossing game, when the platform served her an ad for it in 2016.
According to its 2020 pre-IPO F-1 form, in 2019 DoubleDown LLC,
headquartered in Korea, attracted nearly 2.9 million players per month, driving
more than a quarter-billion dollars in revenue and net income of more than $36
million that year. DoubleDown’s filing goes on to describe its system for getting
the most money possible from its players:
And Facebook openly boasts of its ability to help social casino companies like
DoubleDown find the people who will be most susceptible to spending money on
the game. A case study on Facebook’s promotional site for likely advertising
clients touts the platform’s having delivered 14 percent more app downloads, at a
four times lower cost per customer, for a company called PartyCasino. At a 2014
gaming conference, Julien Codorniou, who directed business for Europe, the
Middle East, and Africa at Facebook at the time, and is now a Facebook vice
president, told an interviewer that the company’s lookalike-audience marketing
product, which finds other Facebook users with interests just like a sample of
your existing customers, is “very good for the gaming companies, because they
can target on Facebook or on mobile specific users or just the whales or just the
customers they want to reactivate… this is our business model.” (Emphasis
mine.) What Codorniou revealed is that Facebook was intimately familiar with
the fact that some people will spend a lot of money, far more than the average
person. Those are the whales, and the platforms help gaming companies find
them. Facebook earns advertising revenue from those companies, as does Google.
According to its policies, Apple takes a 30 percent cut of any money
DoubleDown, Big Fish, or any other social casino company earning more than $1
million a year makes from in-app purchases.
All these platforms have strict rules around real-world gambling, or betting
real money on digital games, as their spokespeople pointed out to me. But the law
doesn’t treat social casino apps as gambling, and neither do the companies’
policies. (None of them would consent to an on-the-record interview about the
issue, and DoubleDown Casino didn’t respond to multiple requests. Big Fish
wrote in a statement that its games are offered for free, for fun, and are not
gambling.) Yet this is a system that is harming people, and the platforms are
crucial to all of this. Without the data available to a game-maker from a platform
like Facebook or Apple or Google, The Loop doesn’t have what it needs to get
going. With it, the cycle of behavior, analysis, winnowed choices, and compulsion
can rev faster and faster.
This vast industry, comprising the game makers and the platforms that feed
them customers, is built to mine our unconscious impulses for money, and the
depth and sophistication of that industry means we as individuals can’t hope to
fight off its influence by ourselves. “I hear it a lot from legislators and the public
saying ‘oh, did these people just not take their statistics, their probability class in
high school? Oh, we need to educate them about how chance works,’” machine-
gambling expert Natasha Dow Schüll says. “This is not about people who really
hope and expect to win. This is about people who discover that this game-playing
is an incredibly efficient vehicle for shifting their internal moods. And you
become very addicted to that, not the prospect of winning.” Any of us, she argues,
can fall into it.
So what should we do? Schüll, like Edelson, says we need to stop measuring
“real-world” losses and start measuring the effects on our decisions. “The real
value that should be measured here is more of the time spent and maybe the
purchases made to continue, rather than anything about gambling or cashing out,”
she says. “As far as legal categories go, I think we don’t have the right category
for assessing the kind of harm that’s done by Candy Crush, by slot machines, and
by Big Fish.”
How would we do that? How would we possibly put a monetary value on the
squishy world of human choice? As the judge wrote in the decision in Machine
Zone’s favor, how do we make this more than “a hodgepodge of hollow claims”?
Happily, we don’t have to build a system to analyze the value of our decisions.
Facebook, among many others, has already done it.
In the early days of the Internet, advertising was sold the same way it was on
television and in magazines. The given unit of measure was cost per thousand, or
CPM. It denoted the number of people a television show or the publisher of
Sports Illustrated could guarantee would see the ad, and the money you paid for
an ad guaranteed you nothing more than the chance to put it in front of a fixed
number of eyeballs.
Facebook, however, has invented an entirely new metric, “cost per action,”
and an entirely new industry, called performance marketing. Rather than
guaranteeing only that your ad will be seen by a certain number of people,
Facebook will instead ask what you want to accomplish with your ad. Do you
want people to buy a brassiere? Subscribe to a newsletter? Join a group? You’ll
pay a rate based on how many people go do the thing you want them to. They
don’t charge by the number of times the ad is seen. They charge by the actions
your customer actually performs. Facebook is so proud of its ability to deliver
outcomes to its advertising customers, rather than just eyeballs, that by 2018, a
senior executive there told me, the majority of its advertising contracts were
written on that basis.
This is a function of scale, of course. When two billion people are spending at
least an hour a day on average on your platform, you can learn enough about them
to deliver the ones that would want to buy a cheese knife or subscribe to a
newsletter about Bitcoin.
But it’s also a function of pattern recognition, and of putting a price on those
patterns. If we wanted to learn the value of a potentially compulsive gambler to a
social casino gaming company, we’d only need to know how much Facebook
charges social casino companies each time a customer downloads an app.
In the 2015 decision against Edelson’s client, the judge wrote that even if he
decided to award damages to her, he’d be in the “unenviable position of pricing
the conversion from virtual gold and chips to virtual wood and rock.” He went on
to write that “such whimsical undertaking may spark the imaginations of children
and ardent game enthusiasts, but it can have no place in federal court.” But the
truth is that Machine Zone, like social casino companies Playtika and
DoubleDown and Big Fish, knows exactly what our time is worth. They price it
out every moment of the day. If we’re looking for numbers to plug into the Hand
Rule—i.e., the probability and gravity of a player losing more money than she can
afford to spend—they’re right there.
SOMEDAY IN THE future, if the world I see coming is actually coming, a writer
will essentially dictate a few thoughts into a piece of software, and it will go on to
write a complete book for her. And the convenience won’t stop there. That author
will also have the option of allowing the software, or some all-encompassing
platform on which she does everything anyway, to look across her
communications and her address book and automatically assemble a heartfelt note
of gratitude to all the people it detects have contributed to her thinking. I find
myself wishing at this moment that such a system existed to assemble for me the
long, complete list of all the people who helped get this book done, because I’m
terrified of forgetting someone. If you’re that person, please accept my gratitude
anyway.
First, I’d like to thank the hundreds of people over the years who have spoken
to me off the record about the places they work, especially those at multinational
corporations. The landscape that journalists and sources at major companies have
to traverse to find one another has on the one hand become logistically easier, in
that I can reach someone in any country in the world in moments. But the
standard information-suppression strategy at your average technology company
involves threatening the careers of those who speak to the press without
permission. And that, combined with a strategy of actively indoctrinating
employees about why their work is on the right side of history (which they
somehow square with forbidding employees from speaking about that work with
outsiders), makes people who will share with me what they know very hard to
find and very important. I believe the future is being determined inside companies
that do not want to reveal their work until it’s too late for the public to have an
effect on it. People at those companies who risk so much to speak with the press
help to rebalance that very lopsided dynamic. I’m enormously grateful to those
who have entrusted me with that responsibility.
This book was made possible by the many friends and colleagues who have
listened to me worry about human nature and technology all these years, and have
helped me shape all of my high hopes and deep anxieties into something
approaching a cogent theory. I took a long walk in a cemetery with the producer
and director Carl Byker in 2015, and began a friendship that took me all over the
world and taught me all about the ancient guidance systems we unconsciously
obey in our lives. His insight, patience, and unrelenting good cheer are a model to
me and got me through this often painful writing process. I’m also very grateful
to producer Kate McMahon, who introduced me to some of the most influential
behavioral scientists in the world.
The Center for Advanced Study in the Behavioral Sciences at Stanford
University, along with the Berggruen Institute, funded a year of writing this book,
and I’m indebted to its director, the political scientist Margaret Levi, for taking a
gamble on me, and to the other fellows for their consideration and generosity with
their time and expertise. The writer and entrepreneur Tim O’Reilly first suggested
I apply to CASBS, and his generous enthusiasm for my subject helped me realize
there might be a book in it.
I continue to be astonished at the good fortune I’ve had to work at NBC News,
where from my first interview I’ve found creative collaborators to help me pursue
my strange, dark subject on television. Noah Oppenheim, Janelle Rodriguez,
Rashida Jones, Libby Leist, Tom Mazzarelli, Elena Nachmanoff, Jessica Kurdali,
Dan Arnall, Betsy Korona, Polly Powell, Audrey Grady, Chiara Sottile, and my
colleagues in the Business, Technology, and Media Unit and at bureaus around
the world have made it possible for me to pursue stories like the ones in these
pages, although on the face of it these sorts of stories offer almost nothing at
which to point a camera. I’m deeply grateful to Heather Allan, who gave me my
first opportunity in television news and taught me from the first day to treat the
technology beat as being about the inequities of capitalism rather than about the
shiny doodads of the future. David Doss showed me how to do the job. And Evan
Groll helped me build the habits of mind and body I needed to be able to put
something like this together.
Popular Science’s editor in chief Mark Jannot gave me the job that set me on
my professional path, and eventually he handed his own job to me, a priceless
gift. I’d also like to thank the many friends and colleagues I found in the
magazine industry, who fought to preserve and improve long-form journalism
even as the business model, gutted by social media, collapsed around them.
I’m grateful to my television agent Peter Goldberg and literary agent Byrd
Leavell for their trust and good counsel, and to my editor, Sam Raim, for shaping
what began as a disjointed series of essays into something resembling a book. The
book’s mistakes are my responsibility. Its strengths are his.
No one understands that work like this is even possible until they’ve seen
someone they admire doing it first, and I’m very lucky to have a family that has
set that example for me. My father, the author Andrew Ward, carved out a very
futuristic career writing about the injustices of his own people decades before it
was fashionable to do so, and raised my sister and me at home while doing it.
Thanks, Dad. My mother, the public health academic Deborah Ward, taught me
about doing the needful, as her mother called it, and helped me understand one of
this book’s major themes, that good decisions usually feel wrong, even when
they’re right. Nathan Ward taught me to find pleasure and dignity in journalism.
Garrett Ward taught me not to be embarrassed by the things I find interesting.
Geoffrey Ward taught me to look for big subjects and great collaborators.
My children, Josephine and Juniper, somehow found it in themselves to
forgive me, at least outwardly, as I repeatedly abandoned them after dinner, on
weekends, and for long stretches of travel for the years it took me to get this done.
My most painful regret in writing this book is the time it took away from them.
The luck I often speak of in describing my time as a journalist—the
opportunity, the privilege, the satisfaction—is all code for the things my wife
Julie makes possible. There is nothing I’ve accomplished that doesn’t directly
depend on her encouragement, love, and sacrifice. I hope that I can someday
support her dreams the way she has supported mine.
Discover Your Next Great Read
Get sneak peeks, book recommendations, and news about your favorite authors.
1. Terrizzi, J.A., Jr., Shook, N.J., McDaniel, M.A. (2013). The behavioral
immune system and social conservatism: A meta-analysis. Evolution and Human
Behavior, 34(2):99–108,
https://linkinghub.elsevier.com/retrieve/pii/S109051381200116X
2. Young, J.L., Hegarty, P. (2019). Reasonable men: Sexual harassment and
norms of conduct in social psychology. Feminism & Psychology, 29(4):453–474.
doi: 10.1177/0959353519855746
3. Gómez, Á., López-Rodríguez, L., Sheikh, H., et al. (2017). The devoted
actor’s will to fight and the spiritual dimension of human conflict. Nature Human
Behavior, 1:673–679. doi: 10.1038/s41562-017-0193-3
4. Hamid, N., Pretus, C., Atran, S., Crockett, M.J., et al. (2019). Neuroimaging
‘will to fight’ for sacred values: An empirical case study with supporters of an Al
Qaeda associate. Royal Society Open Science, Jun 12; 6(6):181585. doi:
10.1098/rsos.181585
CHAPTER 5: GUIDANCE SYSTEMS
1. Mezulis, A.H., Abramson, L.Y., Hyde, J.S., Hankin, B.L. (2004). Is there a
universal positivity bias in attributions? A meta-analytic review of individual,
developmental, and cultural differences in the self-serving attributional bias.
Psychological Bulletin, 130(5):711–747.
2. Waldenström, U., Schytt, E. (2009). A longitudinal study of women’s
memory of labour pain—from 2 months to 5 years after the birth. BJOG, Mar;
116(4):577–583. doi: 10.1111/j.1471-0528.2008.02 020.x. Epub 2008 Dec 9.
PMID: 19076128.
3. Gainsbury, S.M., Russell, A.M.T., King, D.L., Delfabbro, P., Hing, N.
(2016). Migration from social casino games to gambling: Motivations and
characteristics of gamers who gamble. Computers in Human Behavior, 63:59–67.
doi: 10.1016/j.chb.2016.05.021
4. The 192nd Central Court of the Commonwealth of Massachusetts. Session
Laws, Acts (2011), Chapter 194. “An Act Establishing Expanded Gaming in the
Commonwealth.”
https://malegislature.gov/Laws/SessionLaws/Acts/2011/Chapter194
5. Carden, L., Wood, W. (2018). Habit formation and change. Current Opinion
in Behavioral Sciences, 20:117–122. doi: 10.1016/j.cobeha.2017.12.009
CHAPTER 6: LIFE ON THE RAILS
1. Erickson, P., et al. (2015). How Reason Almost Lost Its Mind: The Strange
Career of Cold War Rationality. Chicago: University of Chicago Press.
2. Waytz, A., Morewedge, C.K., Epley, N., Monteleone, G., Gao, J.H.,
Cacioppo, J.T. (2010). Making sense by making sentient: Effectance motivation
increases anthropomorphism. Journal of Personality and Social Psychology,
99(3):410–435. doi: 10.1037/a0020240
3. Risen, J.L. (2016). Believing what we do not believe: Acquiescence to
superstitious beliefs and other powerful intuitions. Psychological Review,
123(2):182–207. doi: 10.1037/rev0000017
4. Weizenbaum, J. (1976). Computer Power and Human Reason: From
Judgment to Calculation. San Francisco: W. H. Freeman and Company.
5. Colby, K.M., Watt, J.B., Gilbert, J.P. (1966). A computer method of
psychotherapy: Preliminary communication. Journal of Nervous and Mental
Disease, 142(2):148–152.
6. Sagan, C. (January 1975). Cosmic calendar. Natural History, 84(10):70–73.
7. ben-Aaron, D. (1985, April 9). Interview: Weizenbaum examines computers
and society. The Tech, http://tech.mit.edu/V105/N16/weisen.16n.html/
8. http://dig.abclocal.go.com/wls/documents/2018/041018-wls-iteam-united-
complaint-doc.pdf
9. “The priority of all other confirmed passengers may be determined based on
a passenger’s fare class, itinerary, status of frequent flyer program membership,
whether the passenger purchased the ticket under select UA corporate travel
agreements, and the time in which the passenger presents him/herself for check-in
without advanced seat assignment.” www.united.com/ual/en/us/fly/contract-of-
carriage.html
CHAPTER 7: WHAT AI ISN’T
1. http://raysolomonoff.com/dartmouth/notebook/dart56ipl_01 17.pdf
2. Rudin, C., Radin, J. (2019). Why are we using black box models in AI
when we don’t need to? A lesson from an explainable AI competition. Harvard
Data Science Review, 1(2). doi: 10.1162/99608f92.5a8a3a3d
3. Fulton, R., Holsinger, B., eds. (2007). History in the Comic Mode: Medieval
Communities and the Matter of Person. New York: Columbia University Press.
4. As summarized by Whittaker, Sir Edmund (1989). A History of the Theories
of Aether and Electricity: Vol. I: The Classical Theories; Vol. II: The Modern
Theories, 1900–1926. Mineola, NY: Dover Publications.
5. O’Hara, J.G., Pricha, W. (1987). Hertz and the Maxwellians: A Study and
Documentation of the Discovery of Electromagnetic Wave Radiation, 1873–1894.
London: P. Peregrinus Limited.
6. Fölsing, A. (1997). Albert Einstein: A Biography. New York: Viking.
7. Kronovet, D. (2017, March 28). Objective functions in machine learning.
Abacus (blog), http://kronosapiens.github.io/blog/2017/03/28/objective-functions-
in-machine-learning.html
8. Patty, J.W., Penn, E.M. (2014). Social Choice and Legitimacy: The
Possibilities of Impossibility. New York: Cambridge University Press. doi:
10.1017/CBO9781139030885
9. Nou, J. (2015). Review of social choice and legitimacy. The New Rambler,
https://newramblerreview.com/images/files/Jennifer_Nou-Review_Patty-and-
Penn.pdf
CHAPTER 9: THE LOOP