Stuart Russell: 3 principles for creating safer AI
斯圖爾特 · 羅素: 人工智慧如何讓我們更美好
Stuart Russell wrote the standard text on AI; now he thinks deeply on AI's future -- and the future of us humans, too. Full bio
Double-click the English transcript below to play the video.
greatest Go players,
頂尖圍棋高手之一,
in Silicon Valley call
我的矽谷朋友們稱之為
a lot faster than we expected.
比我們預期的要快得多。
What about the real world?
那現實世界中情況又如何?
廣闊、複雜得多,
(Decision Problem)。
of the technologies
即將問世的新科技……
is not yet happening in machines,
that the human race has ever written.
更深刻的遠見和洞察力。
further ahead than humans can,
to more information,
in the real world than we can.
做出比人類更好的決策。
everything that we value,
我們所珍視的一切,
to a lot more intelligence,
to what the human race can do.
as some people have described it,
一些人所描述的那樣,
以下的言論呢?
of the human race?
and Stephen Hawking?
史蒂芬霍金的新發明嗎?
has been around for a while.
已經存在很長的時間了。
in a subservient position,
at strategic moments" --
"turning off the power" idea later on --
「關機」這一話題。
feel greatly humbled."
是艾倫 · 圖靈在 1951 年說的。
This is Alan Turing in 1951.
is the father of computer science
the father of AI as well.
他也是人工智慧之父。
比自己更聰明的物種」這個問題時,
more intelligent than your own species,
a few million years ago,
在幾百萬年前就親歷此境,
to discuss whether it was a good idea,
那麽做是不是一個好主意,
they conclude, no,
sadness in their eyes.
看到存亡攸關的憂傷。
something smarter than your own species
並不是什麽妙計」
except stop doing AI,
否則束手無策。
the benefits that I mentioned
to keep doing AI.
the problem a bit more.
可能會是個災難呢?
that the purpose put into the machine
與我們的真正目的相吻合。」
one of the very early learning systems
(Learning System)之後。
better than its creator.
西洋棋下得比發明它的人更好。
I touch to turn to gold,"
所有東西都變成金子。」
that he put into the machine,
and his relatives turned to gold
和家人都變成了金子,
"the King Midas problem"
「邁達斯國王問題」,
which is not, in fact,
不符合實際需求的 「目的」。
「價值取向不一致問題」。
"the value alignment problem."
is not the only part of the problem.
"Fetch the coffee,"
to fetch the coffee?
against interference
所被賦予的任務。」
that I have been given."
極端自我保護的模式在執行,
of an objective that is, in fact,
of the human race --
想要的目標並不一致。
takeaway from this talk.
核心想法,也是價值所在。
the coffee if you're dead.
如果死了,就不能端咖啡了。
Repeat it to yourself three times a day.
每天早晚覆誦三遍。
《2001太空漫步》的劇情。
with the objectives of the humans,
is not superintelligent.
but eventually Dave outwits him
但還是比不過人類戴夫,
pursue objectives.
of altruism, if you like,
of human objectives,
人類的目標、人類的價值。
touchy-feely, goody-goody values.
或者假裝乖巧,
that the human would prefer
無論現狀如何。
its own existence.
its existence whatsoever.
機器對自身生存與否毫不關心。
of humility, if you like.
important to make robots safe.
but it doesn't know what they are.
但卻不知道該價值究竟是什麽。
of single-minded pursuit
而不知變通」的現象。
by observation of human choices,
透過觀察人類的決策,
我們生活的意願,
our lives to be like.
to this question of:
應用到圖靈所說的問題:
as Turing suggested.
right on the back.
going to let you switch it off?
the coffee, I must fetch the coffee,
它會想:「我必須去拿咖啡,
has been listening to my talk,
"I must disable my 'off' switch,
people in Starbucks
星巴克裡干擾我的人都擊暈。」
seems to be inevitable,
a concrete, definite objective.
十分明確的目標。
is uncertain about the objective?
不那麽確定會發生什麽呢?
might switch me off,
principles right there.
the incentive that the robot has
允許讓人類關掉它所包含的動機,
不確定性程度直接相關。
the underlying objective.
it should be pursuing,
what it did wasn't right.
of Greek symbols,
使用些希臘字母,
機器人對人類是絕對有利的。
is provably beneficial to the human.
with a machine that's designed in this way
對我們的生活是是有益的。
but this is the first step
能與人類和諧共處的 AI 的第一步。
with human-compatible AI.
scratching your head over.
這一個原則傷腦筋。
you know, I behave badly.
我行為舉止比較差勁。
and take stuff from the fridge.
從冰箱裡找東西吃,
you don't want the robot doing.
你是不希望機器人去做的。
quite work that way.
is going to copy your behavior.
and maybe help you resist them,
幫助你、制止你。
for any person and for any possible life
可能的生活去預測:
difficulties involved in doing this;
is going to get solved very quickly.
we behave badly.
我們做事不守規矩。
doesn't have to copy the behavior.
未必得要複製那些行為。
any objective of its own.
the desires of one person, the user,
某一個人、某一個用戶的欲望,
the preferences of everybody.
amount of nastiness,
that your nastiness, for example,
你可能會收取賄賂,
and send your kids to school.
但不代表它也會學你偷錢,
it doesn't mean it's going to steal.
send your kids to school.
he took an action that lost the game.
他下錯了那幾步以致輸棋,
through a model of human cognition
limitations -- a very complicated model.
that we can work on understanding.
from my point of view as an AI researcher,
trade off, weigh up the preferences
權衡不同人之間的喜好、需求,
moral philosophers have understood that,
道德哲學家都理解這一點,
looking for collaboration.
when you get that wrong.
走錯了會怎麽樣。
a conversation, for example,
人工智慧助理有這樣的對話,
in a few years' time.
to remind you about dinner tonight."
提醒你別忘了今天的晚宴。」
"What? What dinner?
慶祝結婚 20 周年。」
with the secretary-general at 7:30.
我晚上 7 點半要見秘書長。
my recommendation."
但你沒有理會我的建議。」
I can't just tell him I'm too busy."
我有事,沒空見他。」
for his plane to be delayed."
讓他的航班延誤。」
for lunch tomorrow."
there's a slight mistake going on.
after a hard day's work,
Could you make some dinner?"
who are in more urgent need than you."
比你更緊急,更需要照顧。」
to working on them.
they're going to read everything
閱讀所有人類寫下來的東西。
is human beings doing things
導致其他人對此感到沮喪」。
of data to learn from.
大量的數據中去學習。
strong economic incentive
我們也有足夠的動機去做好這件事。
and the robot has to feed the kids,
機器人得給孩子們做飯,
and there's nothing in the fridge.
但冰箱裡什麽都沒有。
the human value function properly,
貓的情感價值大於其營養價值。
the nutritional value of the cat.
for family dinner."
of the domestic robot industry.
整個居家機器人的產業。
to get this right
我們有足夠的動機把它做對做好。
superintelligent machines.
the definition of AI
beneficial machines.
對我們有益無害的機器人。
about what those objectives are,
that we really want.
we will learn to be better people.
我們也能學會成為更好的人。
because I think they're setting up
seems intuitively really powerful.
這種思想真的很有衝擊力。
this idea that knowledge
and rewriting that programming?
重新編寫自己的程式呢?
it to learn more, as I said,
我們想要它去學習,就像我說的,
as it becomes more correct,
它們才會更明確我們要的東西,
to interpret it correctly.
that books are very biased
判斷出那些富含偏見的書,
和男性精英白人之類的書。
just boil it down to one law,
遠不足以濃縮成一條法則嗎?
a self-driving car
to be able to switch off the car
在汽車運行的過程中將它關閉嗎?
and sensible the person is.
to be switched off.
或者甚至是有惡意的,
random or even malicious,
to be switched off.
figure this out for us.
找到解決的辦法。
十分精彩。
That was amazing.
ABOUT THE SPEAKER
Stuart Russell - AI expertStuart Russell wrote the standard text on AI; now he thinks deeply on AI's future -- and the future of us humans, too.
Why you should listen
Stuart Russell is a professor (and formerly chair) of Electrical Engineering and Computer Sciences at University of California at Berkeley. His book Artificial Intelligence: A Modern Approach (with Peter Norvig) is the standard text in AI; it has been translated into 13 languages and is used in more than 1,300 universities in 118 countries. His research covers a wide range of topics in artificial intelligence including machine learning, probabilistic reasoning, knowledge representation, planning, real-time decision making, multitarget tracking, computer vision, computational physiology, global seismic monitoring and philosophical foundations.
He also works for the United Nations, developing a new global seismic monitoring system for the nuclear-test-ban treaty. His current concerns include the threat of autonomous weapons and the long-term future of artificial intelligence and its relation to humanity.
Stuart Russell | Speaker | TED.com