Google DeepMind CEO Demis Hassabis On ChatGPT, AI, LLMS, and More
Google DeepMind CEO Demis Hassabis On ChatGPT, AI, LLMS, and More
Menu
DECODER
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 1/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
T
oday, I’m talking to Demis Hassabis, the CEO of Google
DeepMind, the newly created division of Google responsible
for AI efforts across the company. Google DeepMind is the
result of an internal merger: Google acquired Demis’ DeepMind
startup in 2014 and ran it as a separate company inside its parent
company, Alphabet, while Google itself had an AI team called Google
Brain.
Google has been showing off AI demos for years now, but with the explosion of
ChatGPT and a renewed threat from Microsoft in search, Google and Alphabet CEO
Sundar Pichai made the decision to bring DeepMind into Google itself earlier this
year to create… Google DeepMind.
What’s interesting is that Google Brain and DeepMind were not necessarily
compatible or even focused on the same things: DeepMind was famous for applying
AI to things like games and protein-folding simulations. The AI that beat world
champions at Go, the ancient board game? That was DeepMind’s AlphaGo.
Meanwhile, Google Brain was more focused on what’s come to be the familiar
generative AI toolset: large language models for chatbots, editing features in Google
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 2/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
Photos, and so on. This was a culture clash and a big structure decision with the goal
of being more competitive and faster to market with AI products.
And the competition isn’t just OpenAI and Microsoft — you might have seen a memo
from a Google engineer floating around the web recently claiming that Google has
no competitive moat in AI because open-source models running on commodity
hardware are rapidly evolving and catching up to the tools run by the giants. Demis
confirmed that the memo was real but said it was part of Google’s debate culture,
and he disagreed with it because he has other ideas about where Google’s
competitive edge might come into play.
Of course, we also talked about AI risk and especially artificial general intelligence.
Demis is not shy that his goal is building an AGI, and we talked through what risks
and regulations should be in place and on what timeline. Demis recently signed onto
a 22-word statement about AI risk with OpenAI’s Sam Altman and others that
simply reads, “Mitigating the risk of extinction from AI should be a global priority
alongside other societal-scale risks such as pandemics and nuclear war.” That’s
pretty chill, but is that the real risk right now? Or is it just a distraction from other
more tangible problems like AI replacing a bunch of labor in various creative
industries? We also talked about the new kinds of labor AI is creating — armies of
low-paid taskers classifying data in countries like Kenya and India in order to train
AI systems. We just published a big feature on these taskers. I wanted to know if
Demis thought these jobs were here to stay or just a temporary side effect of the AI
boom.
This one really hits all the Decoder high points: there’s the big idea of AI, a lot of
problems that come with it, an infinite array of complicated decisions to be made,
and of course, a gigantic org chart decision in the middle of it all. Demis and I got
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 3/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
pretty in the weeds, and I still don’t think we covered it all, so we’ll have to have him
back soon.
Alright, Demis Hassabis, CEO of Google DeepMind. Here we go.
This transcript has been lightly edited for length and clarity
Demis Hassabis, you are the CEO of Google DeepMind. Welcome to Decoder.
I don’t think we have ever had a more perfect Decoder guest. There’s a big idea in
AI. It comes with challenges and problems, and then, with you in particular,
there’s a gigantic org chart move and a set of high-stakes decisions to be made. I
am thrilled that you are here.
Glad to be here.
Let’s start with Google DeepMind itself. Google DeepMind is a new part of Google
that is constructed of two existing parts of Google. There was Google Brain,
which was the AI team we were familiar with as we covered Google that was run
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 4/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
by Jeff Dean. And there was DeepMind, which was your company that you
founded. You sold it to Alphabet in 2014. You were outside of Google. It was run as
a separate company inside that holding company Alphabet structure until just
now. Start at the very beginning. Why were DeepMind and Google Brain separate
to begin with?
As you mentioned, we started DeepMind actually back in 2010, a long time ago now,
especially in the age of AI. So that’s sort of like prehistory. Myself and the co-
founders, we realized coming from academia and seeing what was going on there,
things like deep learning had just been invented. We were big proponents of
reinforcement learning. We could see GPUs and other hardware was coming online,
that a lot of great progress could be made with a focused effort on general learning
systems and also taking some ideas from neuroscience and how the brain works. So
we put all those ingredients together back in 2010. We had this thesis we’d make fast
progress, and that’s what happened with our initial game systems. And then, we
decided in 2014 to join forces with Google at the time because we could see that a lot
more compute was going to be needed. Obviously, Google has the most computers
and had the most computers in the world. That was the obvious home for us to be
able to focus on pushing the research as fast as possible.
So you were acquired by Google, and then somewhere along the way, Google
reoriented itself. They turned into Alphabet, and Google became a division of
Alphabet. There are other divisions of Alphabet, and DeepMind was out of it.
That’s just the part I want to focus on here at the beginning, because there was
what Google was doing with Google Brain, which is a lot of LLM research. I recall,
six years ago, Google was showing off LLMs at Google I/O, but DeepMind was
focused on winning the game [Go] and protein folding, a very different kind of AI
research wholly outside of Google. Why was that outside of Google? Why was that
in Alphabet proper?
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 5/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
Listen to Decoder, a show hosted by The Verge’s Nilay Patel about big ideas — and other
problems. Subscribe here!
That was part of the agreement as we were acquired was that we would pursue
pushing forward research into general AI, or sometimes called AGI, a system that
out of the box can operate across a wide range of cognitive tasks and basically has
all the cognitive capabilities that humans have.
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 6/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
starting, I believe that games was a perfect testing or proving ground for developing
AI algorithms efficiently, quickly, and you can generate a lot of data and the
objective functions are very clear: obviously, winning games or maximizing the
score. There were a lot of reasons to use games in the early days of AI research, and
that was a big part of why we were so successful and why we were able to advance so
quickly with things like AlphaGo, the program that beat the world champion at the
ancient game of Go.
Those were all really important proof points for the whole field really that these
general learning techniques would work. And of course we’ve done a lot of work on
deep learning and neural networks as well. And our specialty, I suppose, was
combining that with reinforcement learning to allow these systems to actively solve
problems and make plans and do things like win games. And in terms of the
differences, we always had that remit to push the research agenda and push things,
advanced science. And that was very much the focus we were given and very much
the focus that I wanted to have. And then, the internal Google AI teams like Google
Brain, they had slightly different remits and were a bit closer to product and
obviously to the rest of Google and infusing Google with amazing AI technology.
And we also had an applied division that was introducing DeepMind technology into
Google products, too. But the cultures were quite different, and the remits were quite
different.
From the outside, the timeline looks like this: everyone’s been working on this for
ages, we’ve all been talking about it for ages. It is a topic of conversation for a
bunch of nerdy journalists like me, a bunch of researchers, we talk about it in the
corner at Google events. Then ChatGPT is released, not even as a product. I don’t
even think Sam [Altman] would call it a great product when it was released, but it
was just released, and people could use it. And everyone freaked out, and
Microsoft releases Bing based on ChatGPT, and the world goes upside down, and
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 7/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
Google reacts by merging DeepMind and Google Brain. That’s what it looks like
from the outside. Is that what it felt like from the inside?
That timeline is correct, but it’s not these direct consequences; it’s more indirect in a
sense. So, Google and Alphabet have always run like this. They let many flowers
bloom, and I think that’s always been the way that even from Larry [Page] and Sergey
[Brin] from the beginning set up Google. And it served them very well, and it’s
allowed them to organically create incredible things and become the amazing
company that it is today. On the research side, I think it’s very compatible with doing
research, which is another reason we chose Google as our partners back in 2014. I
felt they really understood what fundamental and blue sky research was, ambitious
research was, and they were going to facilitate us being and enable us to be super
ambitious with our research. And you’ve seen the results of that, right?
And so, it was surprising, not so much what the technology was because we all
understood that, but the public’s appetite for that and obviously the buzz that
generated. And I think that’s indicative of something we’ve all been feeling for the
last, I would say, two, three years, which is these systems are reaching a level of
maturity now and sophistication where it can really come out of the research phase
and the lab and go into powering incredible next-generation products and
experiences and also breakthroughs, things like AlphaFold directly being useful for
biologists. And so, to me, this is just indicative of a new phase that AI is in of being
practically useful to people in their everyday lives and actually being able to solve
really hard real-world problems that really matter, not just the curiosities or fun, like
games.
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 8/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
When you recognize that shift, then I think that necessitates a change in your
approach as to how you’re approaching the research and how much focus you’re
having on products and those kinds of things. And I think that’s what we all came to
the realization of, which was: now was the time to streamline our AI efforts and
focus them more. And the obvious conclusion of that was to do the merger.
I want to just stop there for one second and ask a philosophical question.
Sure.
It feels like the ChatGPT moment that led to this AI explosion this year was really
rooted in the AI being able to do something that regular people could do. I want
you to write me an email, I want you to write me a screenplay, and maybe the
output of the LLM is a C+, but it’s still something I can do. People can see it. I want
you to fill out the rest of this photo. That’s something people can imagine doing.
Maybe they don’t have the skills to do it, but they can imagine doing it. All the
previous AI demos that we have gotten, even yours, AlphaFold, you’re like, this is
going to model all the proteins in the world.
But I can’t do that; a computer should do that. Even a microbiologist might think,
“That is great. I’m very excited that a computer can do that because I’m just
looking at how much time it would take us, and there’s no way we could ever do
it.” “I want to beat the world champion at Go. I can’t do that. It’s like, fine. A
computer can do that.”
There’s this turn where the computer is starting to do things I can do, and they’re
not even necessarily the most complicated tasks. Read this webpage and deliver a
summary of it to me. But that’s the thing that unlocked everyone’s brain. And I’m
wondering why you think the industry didn’t see that turn coming because we’ve
been very focused on these very difficult things that people couldn’t do, and it
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 9/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
seems like what got everyone is when the computer started doing things people
do all the time.
I think that analysis is correct. I think that is why the large language models have
really entered the public consciousness because it’s something the average person,
that the “Joe Public,” can actually understand and interact with. And, of course,
language is core to human intelligence and our everyday lives. I think that does
explain why chatbots specifically have gone viral in the way they have. Even though I
would say things like AlphaFold, I mean of course I’d be biased in saying this, but I
think it’s actually had the most unequivocally biggest beneficial effects so far in AI
on the world because if you talk to any biologist or there’s a million biologists now,
researchers and medical researchers, have used AlphaFold. I think that’s nearly
every biologist in the world. Every Big Pharma company is using it to advance their
drug discovery programs. I’ve had multiple, dozens, of Nobel Prize-winner-level
biologists and chemists talk to me about how they’re using AlphaFold.
So a certain set of all the world’s scientists, let’s say, they all know AlphaFold, and it’s
affected and massively accelerated their important research work. But of course, the
average person in the street doesn’t know what proteins are even and doesn’t know
what the importance of those things are for things like drug discovery. Whereas
obviously, for a chatbot, everyone can understand, this is incredible. And it’s very
visceral to get it to write you a poem or something that everybody can understand
and process and measure compared to what they do or are able to do.
It seems like that is the focus of productized AI: these chatbot-like interfaces or
these generative products that are going to make stuff for people, and that’s
where the risk has been focused. But even the conversation about risk has
escalated because people can now see, “Oh, these tools can do stuff.” Did you
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 10/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
perceive the same level of scrutiny when you were working on AlphaFold? It
doesn’t seem like anyone thought, “Oh, AlphaFold’s going to destroy humanity.”
No, but there was a lot of scrutiny, but again, it was in a very specialized area, right?
With renowned experts, and actually, we did talk to over 30 experts in the field, from
top biologists to bioethicists to biosecurity people, and actually our partners — we
partnered with the European Bioinformatics Institute to release the AlphaFold
database of all the protein structures, and they guided us as well on how this could
be safely put out there. So there was a lot of scrutiny, and the overwhelming
conclusion from the people we consulted was that the benefits far outweighed any
risks. Although we did make some small adjustments based on their feedback about
which structures to release. But there was a lot of scrutiny, but again, it’s just in a
very expert domain. And just going back to your first question about the generative
models, I do think we are right at the beginning of an incredible new era that’s going
to play out over the next five, 10 years.
Not only in advancing science with AI but in terms of the types of products we can
build to improve people’s everyday lives, billions of people in their everyday lives,
and help them to be more efficient and to enrich their lives. And I think what we’re
seeing today with these chatbots is literally just scratching the surface. There are a
lot more types of AI than generative AI. Generative AI is now the “in” thing, but I
think that planning and deep reinforcement learning and problem-solving and
reasoning, those kinds of capabilities are going to come back in the next wave after
this, along with the current capabilities of the current systems. So I think, in a year
or two’s time, if we were to talk again, we are going to be talking about entirely new
types of products and experiences and services with never-seen-before capabilities.
And I’m very excited about building those things, actually. And that’s one of the
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 11/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
reasons I’m very excited about leading Google DeepMind now in this new era and
focusing on building these AI-powered next-generation products.
Let’s stay in the weeds of Google DeepMind itself, for one more turn. Sundar
Pichai comes to you and says, “All right, I’m the CEO of Alphabet and the CEO of
Google. I can just make this call. I’m going to bring DeepMind into Google, merge
you with Google Brain, you’re going to be the CEO.” How did you react to that
prompt?
It wasn’t like that. It was much more of a conversation between the leaders of the
various different relevant groups and Sundar about pretty much the inflection point
that we’re seeing, the maturity of the systems, what could be possible with those in
the product space, and how to improve experiences for our users, our billions of
users, and how exciting that might be, and what that all requires in totality. Both the
change in focus, a change in the approach to research, the combination of resources
that are required, like compute resources. So there was a big collection of factors to
take into account that we all discussed as a leadership group, and then, conclusions
from that then result in actions, including the merger and also what the plans are
then for the next couple of years and what the focus should be of that merged unit.
Do you perceive a difference being a CEO inside of Google versus being a CEO
inside of Alphabet?
It’s still early days, but I think it’s been pretty similar because, although DeepMind
was an Alphabet company, it was very unusual for another bet, as they call it an
“alpha bet,” which is that we already were very closely integrated and collaborating
with many of the Google product area teams and groups. We had an applied team at
DeepMind whose job it was to translate our research work into features in products
by collaborating with the Google product teams. And so, we’ve had hundreds of
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 12/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
successful launches already actually over the last few years, just quiet ones behind
the scenes. So, in fact, many of the services or devices or systems that you use every
day at Google will have some DeepMind technology under the hood as a component.
So we already had that integrative structure, and then, of course, what we were
famous for was doing the scientific advances and gaming advances, but behind the
scenes, there was a lot of bread and butter work going on that was affecting all parts
of Google.
We were different from other bets where they have to make a business outside of
Google and become an independent business. That was never the goal or the remit
for us, even as an independent bet company. And now, within Google, we’re just more
tightly integrated in terms of the product services, and I see that as an advantage
because we can actually go deeper and do more exciting and ambitious things in
much closer collaboration with these other product teams than we could from
outside of Google. But we still retain some latitude to pick the processes and the
systems that optimize our mission of producing the most capable and general AI
systems in the world.
There’s been reporting that this is actually a culture clash. You’re now in charge
of both. How have you structured the group? How has Google DeepMind
structured under you as CEO, and how are you managing that culture
integration?
Actually, it turns out that the culture’s a lot more similar than perhaps has been
reported externally. And in the end, it’s actually been surprisingly smooth and
pleasant because you’re talking about two world-class research groups, two of the
best AI research organizations in the world, incredible talent on both sides, storied
histories. As we were thinking about the merger and planning it, we were looking at
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 13/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
some document where we listed the top 10 breakthroughs from each group. And
when you take that in totality, it’s like 80–90 percent of over the last decade, of the
breakthroughs that underpin the modern AI industry, from deep reinforcement
learning to transformers, of course. It’s an incredible set of people and talent, and
there’s massive respect for both groups on both sides. And there was actually a lot of
collaboration on a project-based level ongoing over the last decade.
Of course, we all know each other very well. I just think it’s a question of focus and a
bit of coordination across both groups, actually, and more in terms of what are we
going to focus on, other places that it makes sense for the two separate teams to
collaborate on, and maybe de-duplicate some efforts that basically are overlapping.
So fairly obvious stuff, to be honest, but it’s important moving into this new phase
now of where we are into more of an engineering phase of AI, and that requires huge
resources, both compute, engineering, and other things. And, even as a company the
size of Google, we’ve got to pick our bets carefully and be clear about which arrows
we are going to put our wood behind and then focus on those and then massively
deliver on those things. So I think it’s part of the natural course of evolution as to
where we are in the AI journey.
That thing you talked about, “We’re going to combine these groups, we’re going to
pick what we’re doing, we’re going to de-duplicate some efforts.” Those are
structure questions. Have you decided on a structure yet, and what do you think
that structure will be?
The structure’s still evolving. We’re only a couple of months into it. We wanted to
make sure we didn’t break anything, that it was working. Both teams are incredibly
productive, doing super amazing research, but also plugging in to very important
product things that are going on. All of that needs to continue.
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 14/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
You keep saying both teams. Do you think of it as two teams, or are you trying to
make one team?
No, no, for sure it’s one unified team. I like to call it a “super unit,” and I’m very
excited about that. But obviously, we’re still combining that and forming the new
culture and forming the new grouping, including the organizational structures. It’s a
complex thing — putting two big research groups together like this. But I think, by
the end of the summer, we’ll be a single unified entity, and I think that’ll be very
exciting. And we’re already feeling, even a couple of months in, the benefits and the
strengths of that with projects like Gemini that you may have heard of, which is our
next-generation multimodal large models — very, very exciting work going on there,
combining all the best ideas from across both world-class research groups. It’s
pretty impressive to see.
So I think all my training is doing my PhDs and postdocs and so on, obviously I did it
in neuroscience, so I was learning about the brain, but it also taught me how to do
rigorous hypothesis testing and hypothesis generation and then update based on
empirical evidence. The whole scientific method as well as the chess planning, both
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 15/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
can be translated into the business domain. You have to be smart about how to
translate that, you can’t be academic about these things. And often, in the real world,
in business, there’s a lot of uncertainty and hidden information that you don’t know.
So, in chess, obviously all the information’s there for you on the board. You can’t just
directly translate those skills, but I think, in the background, they can be very helpful
if applied in the right way.
There are so many decisions I make every day,it’s hard to come up with one now. But
I tend to try and plan out and scenario a plan many, many years in advance. So I tell
you the way I try to approach things is, I have an end goal. I’m quite good at
imagining things, so that’s a different skill, visualizing or imagining what would a
perfect end state look like, whether that’s organizational or it’s product-based or it’s
research-based. And then, I work back from the end point and then figure out what
all the steps would be required and in what order to make that outcome as likely as
possible.
So that’s a little bit chess-like, right? In the sense of you have some plan that you
would like to get to checkmate your opponent, but you’re many moves away from
that. So what are the incremental things one must do to improve your position in
order to increase the likelihood of that final outcome? And I found that extremely
useful to do that search process from the end goal back to the current state that you
find yourself in.
Let’s put that next to some products. You said there’s a lot of DeepMind
technology and a lot of Google products. The ones that we can all look at are Bard
and then your Search Generative Experience. There’s AI in Google Photos and all
this stuff, but focused on the LLM moment, it’s Bard and the Search Generative
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 16/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
Experience. Those can’t be the end state. They’re not finished. Gemini is coming,
and we’ll probably improve both of those, and all that will happen. When you
think about the end state of those products, what do you see?
The AI systems around Google are also not just in the consumer-facing things but
also under the hood that you may not realize. So even, for example, one of the things
we applied our AI systems to very initially was the cooling systems in Google’s data
centers, enormous data centers, and actually reducing the energy they use by nearly
30 percent that the cooling systems use, which is obviously huge if you multiply that
by all of the data centers and computers they have there. So there are actually a lot
of things under the hood where AI is being used to improve the efficiency of those
systems all the time. But you’re right, the current products are not the end state;
they’re actually just waypoints. And in the case of chatbots and those kinds of
systems, ultimately, they will become these incredible universal personal assistants
that you use multiple times during the day for really useful and helpful things across
your daily lives.
Actually, there’s a whole branch of research going into what’s called tool use. This is
the idea that these large language models or large multimodal models, they’re expert
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 17/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
at language, of course, and maybe a few other capabilities, like math and possibly
coding. But when you ask them to do something specialized, like fold a protein or
play a game of chess or something like this, then actually what they end up doing is
calling a tool, which could be another AI system, that then provides the solution or
the answer to that particular problem. And then that’s transmitted back to the user
via language or pictorially through the central large language model system. So it
may be actually invisible to the user because, to the user, it just looks like one big AI
system that has many capabilities, but under the hood, it could be that actually the
AI system is broken down into smaller ones that have specializations.
And I actually think that probably is going to be the next era. The next generation of
systems will use those kinds of capabilities. And then you can think of the central
system as almost a switch statement that you effectively prompt with language, and
it roots your query or your question or whatever it is you’re asking it to the right tool
to solve that question for you or provide the solution for you. And then transmit that
back in a very understandable way. Again, using through the interface, the best
interface really, of natural language.
Does that process get you closer to an AGI, or does that get you to some
maximum state and you got to do something else?
I think that is on the critical path to AGI, and that’s another reason, by the way, I’m
very excited about this new role and actually doing more products and things
because I actually think the product roadmap from here and the research roadmap
from here toward something like AGI or human-level AI is very complementary. The
kinds of capabilities one would need to push in order to build those kinds of
products that are useful in your everyday life like a universal assistant requires
pushing on some of these capabilities, like planning and memory and reasoning, that
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 18/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
I think are vital for us to get to AGI. So I actually think there’s a really neat feedback
loop now between products and research where they can effectively help each other.
I feel like I had a lot of car CEOs on the show at the beginning of it. I asked all of
them, “When do you think we’re going to get self-driving cars?” And they all said
five years, and they’ve been saying five years for five years, right?
Yes.
I’m going to ask you a version of that question about AGI, but I feel like the
number has gotten smaller recently with people I’ve talked to. How many years
until you think we have AGI?
I think there’s a lot of uncertainty over how many more breakthroughs are required
to get to AGI, big, big breakthroughs — innovative breakthroughs — versus just
scaling up existing solutions. And I think it very much depends on that in terms of
timeframe. Obviously, if there are a lot of breakthroughs still required, those are a lot
harder to do and take a lot longer. But right now, I would not be surprised if we
approached something like AGI or AGI-like in the next decade.
In the next decade. All right, I’m going to come back to you in 10 years. We’re
going to see if that happens.
Sure.
That’s not a straight line, though. You called it the critical path, that’s not a
straight line. There are breakthroughs along the way that might upset the train
and send you along a different path, you think.
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 19/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
and scaling. I wouldn’t be surprised if that happened, either: that we may find that
just scaling the existing systems resulted in diminishing returns in terms of the
performance of the system.
And actually, that would then signal some new innovations were really required to
make further progress. At the moment, I think nobody knows which regime we’re in.
So the answer to that is you have to push on both as hard as possible. So both the
scaling and the engineering of existing systems and existing ideas as well as
investing heavily into exploratory research directions that you think might deliver
innovations that might solve some of the weaknesses in the current systems. And
that’s one advantage of being a large research organization with a lot of resources is
we can bet on both of those things maximally, both of those directions. In a way, I’m
agnostic to that question of “do we need more breakthroughs or will existing
systems just scale all the way?” My view is it’s an empirical question, and one should
push both as hard as possible. And then the results will speak for themselves.
This is a real tension. When you were at DeepMind in Alphabet and you were very
research-focused, and then the research was moved back into Google and
Google’s engineers would turn it into products. And you can see how that
relationship worked. Now, you’re inside of Google. Google is under a lot of
pressure as a company to win this battle. And those are product concerns. Those
are “Make it real for people and go win in the market.” There’s a leaked memo that
went around. It was purportedly from inside Google. It said the company had no
moat and open-source AI models or leaked models would run on people’s
laptops, and they would outpace the company because the history of open
computing would outpace a closed-source competitor. Was that memo real?
is Google, of course, publishing transformers and other things. And TensorFlow and
you look at all the things we’ve done.
We do a huge amount in that space. But I also think there are other considerations
that need to be had as well. Obviously commercial ones but also safety questions
about access to these very powerful systems. What if bad actors can access it? Who
maybe aren’t that technical, so they couldn’t have built it themselves, but they can
certainly reconfigure a system that is out there? What do you do about those things?
And I think that’s been quite theoretical till now, but I think that that is really
important from here all the way to AGI as these systems become more general, more
sophisticated, more powerful. That question is going to be very important about how
does one stop bad actors just using these systems for things they weren’t intended
for but for malicious purposes.
That’s something we need to increasingly come up with, but just back to your
question, look at the history of what Google and DeepMind have done in terms of
coming up with new innovations and breakthroughs and multiple, multiple
breakthroughs over the last decade or more. And I would bet on us, and I’m certainly
very confident that that will continue and actually be even more true over the next
decade in terms of us producing the next key breakthroughs just like we did in the
past.
Do you think that’s the moat: we invented most of this stuff, so we’re going to
invent most of the next stuff?
I don’t really think about it as moats, but I’m an incredibly competitive person. That’s
maybe another thing I got from chess, and many researchers are. Of course, they’re
doing it to discover knowledge, and ultimately, that’s what we are here for is to
improve the human condition. But also, we want to be first to do these things and do
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 21/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
them responsibly and boldly. We have some of the world’s best researchers. I think
we have the biggest collection of great researchers in the world, anywhere in the
world, and an incredible track record. And there’s no reason why that shouldn’t
continue in the future. And in fact, I think with our new organization and
environment might be conducive to even more and faster-paced breakthroughs than
we’ve done in the past.
You’re leading me toward risk and regulation. I want to talk about that, but I want
to start in with just a different spin on it. You’re talking about all the work that
has to be done. You’re talking about deep mind reinforcement learning, how that
works. We ran a gigantic cover story in collaboration with New York Magazine
about the taskers who are actually doing the training, who are actually labeling
the data. There’s a lot of labor conversation with AI along the way. Hollywood
writers are on strike right now because they don’t want ChatGPT to write a bunch
of scripts. I think that’s appropriate.
But then there’s a new class of labor that’s being developed where a bunch of
people around the world are sitting in front of computers and saying, “Yep, that’s
a stop sign. No, that’s not a stop sign. Yep, that’s clothes you can wear. No, that’s
not clothes you can wear.” Is that a forever state? Is that just a new class of work
that needs to be done for these systems to operate? Or does that come to an end?
I think it’s hard to say. I think it’s definitely a moment in time and the current
systems and what they’re requiring at the moment. We’ve been very careful just to
say, from our part, and I think you quoted some of our researchers in that article, to
be very careful to pay living wages and be very responsible about how we do that
kind of work and which partners we use. And we also use internal teams as well. So
actually, I’m very proud of how responsible we’ve been on that type of work. But
going forward, I think there may be ways that these systems, especially once you
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 22/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
have millions and millions of users, effectively can bootstrap themselves. Or one
could imagine AI systems that are capable of actually conversing with themselves or
critiquing themselves.
This would be a bit like turning language systems into a game-like setting, which of
course we’re very expert in and we’ve been thinking about where these
reinforcement learning systems, different versions of them, can actually rate each
other in some way. And it may not be as good as a human rater, but it’s actually a
useful way to do some of the bread and butter rating and then maybe just calibrate it
by checking those ratings with a human rater at the end, rather than getting human
raters to rate everything. So I think there are lots of innovations I can see coming
down the line that will help with this and potentially mean that there’s less
requirement for this all to be done by human raters.
But you think there are always human raters in the mix? Even as you get closer to
AGI, it seems like you need someone to tell the computer if it’s doing a good job or
not.
Let’s take AlphaZero as an example, our general games playing system that ended up
learning, itself, how to play any two-player game, including chess and Go. And it’s
interesting. What happened there is we set up the system so that it could play against
itself tens of millions of times. So, in fact, it built up its own knowledge base. It
started from random, played itself, bootstrapped itself, trained better versions of
itself, and played those off each other in sort of mini-tournaments. But at the end,
you still want to test it against the human world champion or something like this or
an external computer program that was built in a conventional way so that you can
just calibrate your own metrics, which are telling you these systems are improving
according to these objectives or these metrics.
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 23/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
But you don’t know for sure until you calibrate it with an external benchmark or
measure. And depending on what that is, a human rater or human benchmark — a
human expert is often the best thing to calibrate your internal testing against. And
you make sure that your internal tests are actually mapping reality. And again, that’s
something quite exciting about products for researchers because, when you put your
research into products and millions of people are using it every day, that’s when you
get real-world feedback, and there’s no way around that, right? That’s the reality, and
that’s the best test of any theories or any system that you’ve built.
Do you think that work is rewarding or appropriate, the labeling of data for AI
systems? There’s just something about that, which is, “I’m going to tell a
computer how to understand the world so that it might go off in the future and
displace other people.” There’s a loop in there that seems like it’s worth more just
moral or philosophical consideration. Have you spent time thinking about that?
Yeah, I do think about that. I think I don’t really see it like that. I think that what
raters are doing is they’re part of the development cycle of making these systems
safer, more useful for everybody, and more helpful and more reliable. So I think it’s a
critical component. In many industries, we have safety testing of technologies and
products. Today, that’s the best we can do for AI systems is to have human raters. I
think, in the future, the next few years, I think we need a lot more research. And I’ve
been calling for this, and we are doing this ourselves, but it needs more than just one
organization to do this, is great, robust evaluation benchmarks for capabilities so
that we know if a system passes these benchmarks, then it has certain properties,
and it’s safe and it’s reliable in these particular ways.
And right now, I think we are in the space of many researchers in academia and civil
society and elsewhere, we have a lot of good suggestions for what those tests could
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 24/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
be, but I don’t think they are robust or practical yet. I think they’re basically
theoretical and philosophical in nature, and I think they need to be made practical so
that we can measure our systems empirically against those tests and then that gives
us some assurances about how the system will perform. And I think once we have
those, then the need for this human rating testing feedback will be reduced. I just
think that’s required in the volumes that’s required now because we don’t have these
kinds of independent benchmarks yet. Partly because we haven’t rigorously defined
what those properties are. I mean, it’s almost a neuroscience and psychology and
philosophy area as well, right? A lot of these terms have not been defined properly,
even for the human brain.
You’ve signed a letter from the Center for AI Safety — OpenAI’s Sam Altman and
others have also signed this letter — that warns against the risk from AI. And yet,
you’re pushing on, Google’s in the market, you’ve got to win, you’ve described
yourself as competitive. There’s a tension there: needing to win in the market
with products and “Oh boy, please regulate us because raw capitalism will drive
us off the cliff with AI if we don’t stop it in some way.” How do you balance that
risk?
Try and anticipate what the issues might be if one was successful ahead of time. Not
in hindsight, and perhaps this happened with social media, for example, where it is
this incredible growth story. Obviously, it’s done a lot of good in the world, but then
it turns out 15 years later we realize there are some unintended consequences as well
to those types of systems. And I would like to chart a different path with AI. And I
think it’s such a profound and important and powerful technology. I think we have to
do that with something as potentially as transformative as AI. And it doesn’t mean
no mistakes will be made. It’s very new, anything new, you can’t predict everything
ahead of time, but I think we can try and do the best job we can.
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 25/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
And that includes mostly, at this stage, doing more research into these areas, like
coming up with better evaluations and benchmarks to rigorously test the
capabilities of these frontier systems.
You talked about tool usage for AI models, you ask an LLM to do something, it
goes off and asks AlphaFold to fold the protein for you. Combining systems like
that, integrating systems like that, historically that’s where emergent behaviors
appear, things you couldn’t have predicted start happening. Are you worried
about that? There’s not a rigorous way to test that.
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 26/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
Right, exactly. I think that’s exactly the sort of thing we should be researching and
thinking about ahead of time is: as tool use becomes more sophisticated and you can
combine different AI systems together in different ways, there is scope for emergent
behavior. Of course, that emergent behavior may be very desirable and be extremely
useful, but it could also potentially be harmful in the wrong hands and in the hands
of bad actors, whether that’s individuals or even nation-states.
Let’s say the United States and the EU and China all agree on some framework to
regulate AI, and then North Korea or Iran says, “Fuck it, no rules.” And that
becomes a center of bad actor AI research. How does that play out? Do you
foresee a world in which that’s possible?
Yeah, I think that is a possible world. This is why I’ve been talking to governments —
UK, US mostly, but also EU — on I think whatever regulations or guardrails or
whatever that is that transpires over the next few years, and tests. They ideally would
be international, and there would be international cooperation around those
safeguards and international agreement around deployment of these systems and
other things. Now, I don’t know how likely that is given the geopolitical tensions
around the world, but that is by far the best state. And I think what we should be
aiming for if we can.
If the government here passes a rule. It says, “Here’s what Google is allowed to do,
here’s what Microsoft is allowed to do. You are in charge, you are accountable.”
And you can go say, “All right, we’re just not running this code in our data center.
We are not going to have these capabilities; it’s not legal.” If I’m just a person with
a MacBook, would you accept some limitation on what a MacBook could do
because the threat from AI is so scary? That’s the thing I worry about. Practically,
if you have open-source models and people are going to use them for weird
things, are we going to tell Intel to restrict what its chips can do? How would we
implement that such that it actually affects everyone? And not just, we’re going to
throw Demis in jail if Google does stuff we don’t like.
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 27/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
I think those are the big questions that are being debated right now. And I do worry
about that. On the one hand, there are a lot of benefits of open-sourcing and
accelerating scientific discourse and lots of advances happen there and it gives
access to many developers. On the other hand, there could be some negative
consequences with that if there are bad individual actors that do bad things with
that access and that proliferates. And I think that’s a question for the next few years
that will need to be resolved. Because right now, I think it’s okay because the systems
are not that sophisticated or that powerful and therefore not that risky.
But I think, as systems increase in their power and generality, the access question
will need to be thought about from government and how they want to restrict that or
control that or monitor that is going to be an important question. I don’t have any
answers for you because I think this is a societal question actually that requires
stakeholders from right across society to come together and weigh up the benefits
with the risks there.
Google’s own work, you said we’re not there yet, but Google’s own work in AI
certainly had some controversy associated with this around responsibility,
around what the models can do or can’t do. There’s a famous “Stochastic Parrots”
paper from Emily Bender and Timnit Gebru and Margaret Mitchell that led to a
lot of controversy inside of Google. It led to them leaving. Did you read that paper
and think, “Okay, this is correct. LLMs are going to lie to people and Google will
be responsible for that”? And how do you think about that now with all of the
scrutiny?
Yeah, look, the large language models, and I think this is one reason that Google’s
been very responsible with this, is that we know that they hallucinate and they can
be inaccurate. And that’s one of the key areas that has to be improved over the next
few years is factuality and grounding and making sure that they don’t spread
disinformation, these kinds of things. And that’s very much top of mind for us. And
we have many ideas of how to improve that. And our old DeepMind’s Sparrow
language model, which we published a couple of years ago, was an experiment into
just how good can we get factuality and rules adherence in these systems. And turns
out, we can maybe make it an order of magnitude better, but it sometimes comes at
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 28/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
the expense of lucidness or creativity on the part of the language model and
therefore usefulness.
So it’s a bit of a Pareto frontier where, if you improve one dimension, you reduce the
capability in another dimension. And ideally, what we want to do in the next phases
and the next generations of systems is combine the best of both worlds — keep the
creativity and lucidness and funness of the current systems but improve their
factuality and reliability. And we’ve got a long way to go on that. But I can see things
improving, and I don’t see any theoretical reason why these systems can’t get to
extremely high levels of accuracy and reliability in the next few years.
When you’re using the Google Search Generative Experience, do you believe what
it says?
I do. I sometimes double-check things, especially in the scientific domain where I’ve
had very funny situations where, actually all of these models do this, where you ask
them to summarize an area of research, which I think would be super useful if they
could do that, and then say, “Well, what are the key papers I should read?” And they
come up with very plausible sounding papers with very plausible author lists. But
then, when you go and look into it, it turns out that they’re just like the most famous
people in that field or the titles from two different papers combined together. But of
course, they’re extremely plausible as a collection of words. And I think, there what
needs to happen is these systems need to understand that citations and papers and
author lists are a unitary block rather than a word-by-word prediction.
There are interesting cases like that where we need to improve, and there’s
something which is, of course, us as wanting to advance the frontiers of science,
that’s a particularly interesting use case that we would like to improve and fix — for
our own needs as well. I’d love these systems to better summarize for me “here are
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 29/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
the top five papers to read about a particular disease” or something like that to just
quickly onboard you in that particular area. I think it would be incredibly useful.
I’ll tell you, I googled my friend John Gruber, and SGE confidently told me that he
pioneered the use of a Mac in newspapers and invented WebKit. I don’t know
where that came from. Is there a quality level, a truthfulness level that you need to
hit before you roll that out to the mass audience?
Yeah, we think about this all the time, especially at Google because of the incredibly
high standards Google holds itself to on things like search and that we all rely on
every day and every moment of every day, really, and we want to get toward that level
of reliability. Obviously, we’re a long, long, long way away from that at the moment
with not just us but anybody with their generative systems. But that’s the gold
standard. And actually, things like tool use can come in very handy here where you
could, in effect, build these systems so that they fact-check themselves, perhaps even
using search or other reliable sources, cross-reference, just like a good researcher
would, cross-reference your facts. Also having a better understanding of the world.
What are research papers? What entities are they?
So these systems need to have a better understanding of the media they’re dealing
with. And maybe also give these systems the ability to reason and plan because then
they could potentially turn that on their own outputs and critique themselves. And
again, this is something we have a lot of experience in in games programs. They
don’t just output the first move that you think of in chess or Go. You actually plan
and do some search around that and then back up. And sometimes they change their
minds and switch to a better move. And you could imagine some process like that
with words and language as well.
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 30/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
There’s the concept of model collapse. That we’re going to train LLMs on LLM-
generated data, and that’s going to go into a circle. When you talk about cross-
referencing facts, and I think about Google — Google going out in the web and
trying to cross-reference a bunch of stuff but maybe all that stuff has been
generated by LLMs that were hallucinating in 2023. How do you guard against
that?
We are working on some pretty cool solutions to that. I think the answer is, and this
is an answer to deepfakes as well, is to do some encrypted watermarking,
sophisticated watermarking, that can’t be removed easily or at all, and it’s probably
built into the generative models themselves, so it’s part of the generative process. We
hope to release that and maybe provide it to third parties as well as a generic
solution. But I think that the industry in the field needs those types of solutions
where we can mark generated media, be that images, audio, perhaps even text with
some Kitemark that says to the user and future AI systems that these were AI-
generated. And I think that’s a very, very pressing need right now for near-term
issues with AI like deepfakes and disinformation and so on. But I actually think a
solution is on the horizon now.
I had Microsoft CTO and EVP of AI Kevin Scott on the show a few weeks ago. He
said something very similar. I promised him that we would do a one-hour episode
on metadata. So you’re coming for that one. If I know this audience, a full hour on
metadata ideas will be our most popular episode ever.
Demis, thank you so much for coming on Decoder. You have to come back soon.
Thanks so much.
Decoder with Nilay Patel / A podcast about big ideas and other problems
SUBSCRIBE NOW!
5 COMMENTS (5 NEW)
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 31/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
Google’s first truly custom chip for Pixel devices won’t arrive
until 2025
Nintendo’s CEO says its online accounts are the key to its
next console transition
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 32/33
7/11/23, 7:08 AM Google DeepMind CEO Demis Hassabis on ChatGPT, AI, LLMs, and more - The Verge
TERMS OF USE / PRIVACY NOTICE / COOKIE POL ICY / DO NOT SEL L OR SH ARE MY PERSON AL INFO / L ICENSING FAQ
/ ACCESSIBIL ITY / PL ATFORM STATUS / HOW WE RATE AND REVIEW PRODUCTS
© 2 02 3 VOX M E D I A , L L C . A L L R I G H T S R E S E RV E D
https://www.theverge.com/23778745/demis-hassabis-google-deepmind-ai-alphafold-risks 33/33