AI isn’t as smart as you think – but it could be
Link: https://www.ted.com/talks/jeff_dean_ai_isn_t_as_smart_as_you_think_but_it_could_be
Speaker: Jeff Dean
Jeffrey Adgate “Jeff” Dean (born July 23, 1968) is an American computer scientist and software engineer. Since 2018, he has been the lead of Google AI.[1] He was appointed Alphabet’s chief scientist in 2023 after a reorganization of Alphabet’s AI focused groups.[2]
Date: August 2021
文章目录
- AI isn't as smart as you think -- but it could be
- Introduction
- Vocabulary
- Transcript
- Q&A with Chris Anderson
- Summary
- 后记
Introduction
What is AI, really? Jeff Dean, the head of Google’s AI efforts, explains the underlying technology that enables artificial intelligence to do all sorts of things, from understanding language to diagnosing disease – and presents a roadmap for building better, more responsible systems that have a deeper understanding of the world. (Followed by a Q&A with head of TED Chris Anderson)
Vocabulary
wedge into:挤进
when we were all wedged into a tiny office space 当我们都挤在一个狭小的办公室里
emulate:美 [ˈemjuleɪt] 模仿,仿效
a series of interconnected artificial neurons that loosely emulate the properties of your real neurons. 一系列相互连接的人造神经元,松散地模拟了真实神经元的特性。
tailored to:合身的,适合的,裁剪的
we also got excited about how we could build hardware that was better tailored to the kinds of computations neural networks wanted to do. 我们还对如何构建更适合神经网络想要进行的计算类型的硬件感到兴奋。
hone:磨砺,磨练
hydroponic:美 [ˌhaɪdrə’pɒnɪk] 水耕法的, 水栽法的
I’ve been honing my gardening skills, experimenting with vertical hydroponic gardening. 我一直在磨练我的园艺技能,尝试垂直水培园艺。
Transcript
Hi, I’m Jeff.
I lead AI Research and Health at Google.
I joined Google more than 20 years ago,
when we were all wedged
into a tiny office space,
above what’s now a T-Mobile store
in downtown Palo Alto.
I’ve seen a lot of computing
transformations in that time,
and in the last decade, we’ve seen AI
be able to do tremendous things.
But we’re still doing it
all wrong in many ways.
That’s what I want
to talk to you about today.
But first, let’s talk
about what AI can do.
So in the last decade,
we’ve seen tremendous progress
in how AI can help computers see,
understand language,
understand speech better than ever before.
Things that we couldn’t do
before, now we can do.
If you think about computer vision alone,
just in the last 10 years,
computers have effectively
developed the ability to see;
10 years ago, they couldn’t see,
now they can see.
You can imagine this has had
a transformative effect
on what we can do with computers.
So let’s look at a couple
of the great applications
enabled by these capabilities.
We can better predict flooding,
keep everyone safe,
using machine learning.
We can translate over 100 languages
so we all can communicate better,
and better predict and diagnose disease,
where everyone gets
the treatment that they need.
So let’s look at two key components
that underlie the progress
in AI systems today.
The first is neural networks,
a breakthrough approach to solving
some of these difficult problems
that has really shone
in the last 15 years.
But they’re not a new idea.
And the second is computational power.
It actually takes a lot
of computational power
to make neural networks
able to really sing,
and in the last 15 years,
we’ve been able to have that,
and that’s partly what’s enabled
all this progress.
But at the same time,
I think we’re doing several things wrong,
and that’s what I want
to talk to you about
at the end of the talk.
First, a bit of a history lesson.
So for decades,
almost since the very
beginning of computing,
people have wanted
to be able to build computers
that could see, understand language,
understand speech.
The earliest approaches
to this, generally,
people were trying to hand-code
all the algorithms
that you need to accomplish
those difficult tasks,
and it just turned out
to not work very well.
But in the last 15 years,
a single approach
unexpectedly advanced all these different
problem spaces all at once:
neural networks.
So neural networks are not a new idea.
They’re kind of loosely based
on some of the properties
that are in real neural systems.
And many of the ideas
behind neural networks
have been around since the 1960s and 70s.
A neural network is what it sounds like,
a series of interconnected
artificial neurons
that loosely emulate the properties
of your real neurons.
An individual neuron
in one of these systems
has a set of inputs,
each with an associated weight,
and the output of a neuron
is a function of those inputs
multiplied by those weights.
So pretty simple,
and lots and lots of these work together
to learn complicated things.
So how do we actually learn
in a neural network?
It turns out the learning process
consists of repeatedly making
tiny little adjustments
to the weight values,
strengthening the influence
of some things,
weakening the influence of others.
By driving the overall system
towards desired behaviors,
these systems can be trained
to do really complicated things,
like translate
from one language to another,
detect what kind
of objects are in a photo,
all kinds of complicated things.
I first got interested in neural networks
when I took a class on them
as an undergraduate in 1990.
At that time,
neural networks showed
impressive results on tiny problems,
but they really couldn’t scale to do
real-world important tasks.
But I was super excited.
(Laughter)
I felt maybe we just needed
more compute power.
And the University of Minnesota
had a 32-processor machine.
I thought, "With more compute power,
boy, we could really make
neural networks really sing."
So I decided to do a senior thesis
on parallel training of neural networks,
the idea of using processors in a computer
or in a computer system
to all work toward the same task,
that of training neural networks.
32 processors, wow,
we’ve got to be able
to do great things with this.
But I was wrong.
Turns out we needed about a million times
as much computational power
as we had in 1990
before we could actually get
neural networks to do impressive things.
But starting around 2005,
thanks to the computing progress
of Moore’s law,
we actually started to have
that much computing power,
and researchers in a few universities
around the world started to see success
in using neural networks for a wide
variety of different kinds of tasks.
I and a few others at Google
heard about some of these successes,
and we decided to start a project
to train very large neural networks.
One system that we trained,
we trained with 10 million
randomly selected frames
from YouTube videos.
The system developed the capability
to recognize all kinds
of different objects.
And it being YouTube, of course,
it developed the ability
to recognize cats.
YouTube is full of cats.
(Laughter)
But what made that so remarkable
is that the system was never told
what a cat was.
So using just patterns in data,
the system honed in on the concept
of a cat all on its own.
All of this occurred at the beginning
of a decade-long string of successes,
of using neural networks
for a huge variety of tasks,
at Google and elsewhere.
Many of the things you use every day,
things like better speech
recognition for your phone,
improved understanding
of queries and documents
for better search quality,
better understanding of geographic
information to improve maps,
and so on.
Around that time,
we also got excited about how we could
build hardware that was better tailored
to the kinds of computations
neural networks wanted to do.
Neural network computations
have two special properties.
The first is they’re very tolerant
of reduced precision.
Couple of significant digits,
you don’t need six or seven.
And the second is that all the
algorithms are generally composed
of different sequences
of matrix and vector operations.
So if you can build a computer
that is really good at low-precision
matrix and vector operations
but can’t do much else,
that’s going to be great
for neural-network computation,
even though you can’t use it
for a lot of other things.
And if you build such things,
people will find amazing uses for them.
This is the first one we built, TPU v1.
“TPU” stands for Tensor Processing Unit.
These have been used for many years
behind every Google search,
for translation,
in the DeepMind AlphaGo matches,
so Lee Sedol and Ke Jie
maybe didn’t realize,
but they were competing
against racks of TPU cards.
And we’ve built a bunch
of subsequent versions of TPUs
that are even better and more exciting.
But despite all these successes,
I think we’re still doing
many things wrong,
and I’ll tell you about three
key things we’re doing wrong,
and how we’ll fix them.
The first is that most
neural networks today
are trained to do one thing,
and one thing only.
You train it for a particular task
that you might care deeply about,
but it’s a pretty heavyweight activity.
You need to curate a data set,
you need to decide
what network architecture you’ll use
for this problem,
you need to initialize the weights
with random values,
apply lots of computation
to make adjustments to the weights.
And at the end, if you’re lucky,
you end up with a model
that is really good
at that task you care about.
But if you do this over and over,
you end up with thousands
of separate models,
each perhaps very capable,
but separate for all the different
tasks you care about.
But think about how people learn.
In the last year, many of us
have picked up a bunch of new skills.
I’ve been honing my gardening skills,
experimenting with vertical
hydroponic gardening.
To do that, I didn’t need to relearn
everything I already knew about plants.
I was able to know
how to put a plant in a hole,
how to pour water, that plants need sun,
and leverage that
in learning this new skill.
Computers can work
the same way, but they don’t today.
If you train a neural
network from scratch,
it’s effectively like forgetting
your entire education
every time you try to do something new.
That’s crazy, right?
So instead, I think we can
and should be training
multitask models that can do
thousands or millions of different tasks.
Each part of that model would specialize
in different kinds of things.
And then, if we have a model
that can do a thousand things,
and the thousand and first
thing comes along,
we can leverage
the expertise we already have
in the related kinds of things
so that we can more quickly be able
to do this new task,
just like you, if you’re confronted
with some new problem,
you quickly identify
the 17 things you already know
that are going to be helpful
in solving that problem.
Second problem is that most
of our models today
deal with only a single
modality of data –
with images, or text or speech,
but not all of these all at once.
But think about how you
go about the world.
You’re continuously using all your senses
to learn from, react to,
figure out what actions
you want to take in the world.
Makes a lot more sense to do that,
and we can build models in the same way.
We can build models that take in
these different modalities of input data,
text, images, speech,
but then fuse them together,
so that regardless of whether the model
sees the word “leopard,”
sees a video of a leopard
or hears someone say the word “leopard,”
the same response
is triggered inside the model:
the concept of a leopard
can deal with different
kinds of input data,
even nonhuman inputs,
like genetic sequences,
3D clouds of points,
as well as images, text and video.
The third problem
is that today’s models are dense.
There’s a single model,
the model is fully activated
for every task,
for every example
that we want to accomplish,
whether that’s a really simple
or a really complicated thing.
This, too, is unlike
how our own brains work.
Different parts of our brains
are good at different things,
and we’re continuously calling
upon the pieces of them
that are relevant for the task at hand.
For example, nervously watching
a garbage truck
back up towards your car,
the part of your brain that thinks
about Shakespearean sonnets
is probably inactive.
(Laughter)
AI models can work the same way.
Instead of a dense model,
we can have one
that is sparsely activated.
So for particular different tasks,
we call upon different parts of the model.
During training, the model can also learn
which parts are good at which things,
to continuously identify what parts
it wants to call upon
in order to accomplish a new task.
The advantage of this is we can have
a very high-capacity model,
but it’s very efficient,
because we’re only calling
upon the parts that we need
for any given task.
So fixing these three things, I think,
will lead to a more powerful AI system:
instead of thousands of separate models,
train a handful of general-purpose models
that can do thousands
or millions of things.
Instead of dealing with single modalities,
deal with all modalities,
and be able to fuse them together.
And instead of dense models,
use sparse, high-capacity models,
where we call upon the relevant
bits as we need them.
We’ve been building a system
that enables these kinds of approaches,
and we’ve been calling
the system “Pathways.”
So the idea is this model
will be able to do
thousands or millions of different tasks,
and then, we can incrementally
add new tasks,
and it can deal
with all modalities at once,
and then incrementally learn
new tasks as needed
and call upon the relevant
bits of the model
for different examples or tasks.
And we’re pretty excited about this,
we think this is going
to be a step forward
in how we build AI systems.
But I also wanted
to touch on responsible AI.
We clearly need to make sure
that this vision of powerful AI systems
benefits everyone.
These kinds of models raise
important new questions
about how do we build them with fairness,
interpretability, privacy and security,
for all users in mind.
For example, if we’re going
to train these models
on thousands or millions of tasks,
we’ll need to be able to train them
on large amounts of data.
And we need to make sure that data
is thoughtfully collected
and is representative of different
communities and situations
all around the world.
And data concerns are only
one aspect of responsible AI.
We have a lot of work to do here.
So in 2018, Google published
this set of AI principles
by which we think about developing
these kinds of technology.
And these have helped guide us
in how we do research in this space,
how we use AI in our products.
And I think it’s a really helpful
and important framing
for how to think about these deep
and complex questions
about how we should
be using AI in society.
We continue to update these
as we learn more.
Many of these kinds of principles
are active areas of research –
super important area.
Moving from single-purpose systems
that kind of recognize patterns in data
to these kinds of general-purpose
intelligent systems
that have a deeper
understanding of the world
will really enable us to tackle
some of the greatest problems
humanity faces.
For example,
we’ll be able to diagnose more disease;
we’ll be able to engineer better medicines
by infusing these models
with knowledge of chemistry and physics;
we’ll be able to advance
educational systems
by providing more individualized tutoring
to help people learn
in new and better ways;
we’ll be able to tackle
really complicated issues,
like climate change,
and perhaps engineering
of clean energy solutions.
So really, all of these kinds of systems
are going to be requiring
the multidisciplinary expertise
of people all over the world.
So connecting AI
with whatever field you are in,
in order to make progress.
So I’ve seen a lot
of advances in computing,
and how computing, over the past decades,
has really helped millions of people
better understand the world around them.
And AI today has the potential
to help billions of people.
We truly live in exciting times.
Thank you.
(Applause)
Q&A with Chris Anderson
Chris Anderson: Thank you so much.
I want to follow up on a couple things.
This is what I heard.
Most people’s traditional picture of AI
is that computers recognize
a pattern of information,
and with a bit of machine learning,
they can get really good at that,
better than humans.
What you’re saying is those patterns
are no longer the atoms
that AI is working with,
that it’s much richer-layered concepts
that can include all manners
of types of things
that go to make up a leopard, for example.
So what could that lead to?
Give me an example
of when that AI is working,
what do you picture happening in the world
in the next five or 10 years
that excites you?
Jeff Dean: I think
the grand challenge in AI
is how do you generalize
from a set of tasks
you already know how to do
to new tasks,
as easily and effortlessly as possible.
And the current approach of training
separate models for everything
means you need lots of data
about that particular problem,
because you’re effectively trying
to learn everything
about the world
and that problem, from nothing.
But if you can build these systems
that already are infused with how to do
thousands and millions of tasks,
then you can effectively
teach them to do a new thing
with relatively few examples.
So I think that’s the real hope,
that you could then have a system
where you just give it five examples
of something you care about,
and it learns to do that new task.
CA: You can do a form
of self-supervised learning
that is based on remarkably
little seeding.
JD: Yeah, as opposed to needing
10,000 or 100,000 examples
to figure everything in the world out.
CA: Aren’t there kind of terrifying
unintended consequences
possible, from that?
JD: I think it depends
on how you apply these systems.
It’s very clear that AI
can be a powerful system for good,
or if you apply it in ways
that are not so great,
it can be a negative consequence.
So I think that’s why it’s important
to have a set of principles
by which you look at potential uses of AI
and really are careful and thoughtful
about how you consider applications.
CA: One of the things
people worry most about
is that, if AI is so good at learning
from the world as it is,
it’s going to carry forward
into the future
aspects of the world as it is
that actually aren’t right, right now.
And there’s obviously been
a huge controversy about that
recently at Google.
Some of those principles
of AI development,
you’ve been challenged that you’re not
actually holding to them.
Not really interested to hear
about comments on a specific case,
but … are you really committed?
How do we know that you are
committed to these principles?
Is that just PR, or is that real,
at the heart of your day-to-day?
JD: No, that is absolutely real.
Like, we have literally hundreds of people
working on many of these
related research issues,
because many of those
things are research topics
in their own right.
How do you take data from the real world,
that is the world as it is,
not as we would like it to be,
and how do you then use that
to train a machine-learning model
and adapt the data bit of the scene
or augment the data with additional data
so that it can better reflect
the values we want the system to have,
not the values that it sees in the world?
CA: But you work for Google,
Google is funding the research.
How do we know that the main values
that this AI will build
are for the world,
and not, for example, to maximize
the profitability of an ad model?
When you know everything
there is to know about human attention,
you’re going to know so much
about the little wriggly,
weird, dark parts of us.
In your group, are there rules
about how you hold off,
church-state wall
between a sort of commercial push,
“You must do it for this purpose,”
so that you can inspire
your engineers and so forth,
to do this for the world, for all of us.
JD: Yeah, our research group
does collaborate
with a number of groups across Google,
including the Ads group,
the Search group, the Maps group,
so we do have some collaboration,
but also a lot of basic research
that we publish openly.
We’ve published more
than 1,000 papers last year
in different topics,
including the ones you discussed,
about fairness, interpretability
of the machine-learning models,
things that are super important,
and we need to advance
the state of the art in this
in order to continue to make progress
to make sure these models
are developed safely and responsibly.
CA: It feels like we’re at a time
when people are concerned
about the power of the big tech companies,
and it’s almost, if there was ever
a moment to really show the world
that this is being done
to make a better future,
that is actually key to Google’s future,
as well as all of ours.
JD: Indeed.
CA: It’s very good to hear you
come and say that, Jeff.
Thank you so much for coming here to TED.
JD: Thank you.
(Applause)
Summary
Jeff Dean’s speech focuses on the transformative potential of AI, highlighting the significant progress made in the last decade. He discusses the capabilities of AI in various domains, such as computer vision, language understanding, and medical diagnosis. Dean attributes this progress to advancements in neural networks and computational power. However, he acknowledges several shortcomings in current AI systems, including their single-task nature, limited modalities, and dense architecture.
Dean proposes solutions to address these shortcomings, advocating for the development of multitask models capable of handling diverse tasks and modalities. He emphasizes the importance of sparse, high-capacity models that can efficiently activate relevant components for specific tasks. Dean introduces the concept of “Pathways,” a system designed to enable these approaches, which he believes will represent a significant step forward in AI research and development.
In addition to discussing technical advancements, Dean emphasizes the importance of responsible AI, highlighting the need for fairness, interpretability, privacy, and security in AI systems. He outlines Google’s AI principles as a guiding framework for responsible AI development and acknowledges the ongoing research and challenges in this area. Dean concludes his speech by expressing optimism about the potential of AI to address complex societal challenges and the importance of multidisciplinary collaboration in realizing this potential.
后记
2024年4月29日18点23分完成这篇演讲的学习。