Placeholder Image

Subtitles section Play video

  • This is Lee Sedol.

    譯者: 易帆 余 審譯者: Wilde Luo

  • Lee Sedol is one of the world's greatest Go players,

    這是李世石。

  • and he's having what my friends in Silicon Valley call

    李世石是全世界 頂尖圍棋高手之一,

  • a "Holy Cow" moment --

    此時,他正在經歷的是 我的矽谷朋友們稱之為

  • (Laughter)

    「我的媽呀!」的時刻......

  • a moment where we realize

    (笑聲)

  • that AI is actually progressing a lot faster than we expected.

    在這一刻讓我們意識到,

  • So humans have lost on the Go board. What about the real world?

    原來人工智慧發展的進程 比我們預期的要快得多。

  • Well, the real world is much bigger,

    人類已在圍棋博弈中落敗, 那現實世界中情況又如何?

  • much more complicated than the Go board.

    當然啦,現實世界要比棋盤 廣闊、複雜得多,

  • It's a lot less visible,

    它也遠不如棋盤上那麽黑白分明,

  • but it's still a decision problem.

    但仍然是個判定問題 (Decision Problem)。

  • And if we think about some of the technologies

    如果我們思考一些 即將問世的新科技……

  • that are coming down the pike ...

    新井紀子提到機器仍無法 「閱讀」,

  • Noriko [Arai] mentioned that reading is not yet happening in machines,

    至少無法真正理解文本含義。

  • at least with understanding.

    但這項能力最終會被機器掌握,

  • But that will happen,

    而當這一切發生時,

  • and when that happens,

    不久之後,

  • very soon afterwards,

    機器就能讀遍所有人類寫下的東西。

  • machines will have read everything that the human race has ever written.

    這會讓機器擁有比人類 更深刻的遠見和洞察力。

  • And that will enable machines,

    就如我們在這場圍棋博弈中所見,

  • along with the ability to look further ahead than humans can,

    如果機器能接觸到比人類更多的信息,

  • as we've already seen in Go,

    那機器將能夠在現實世界中 做出比人類更好的決策。

  • if they also have access to more information,

    那這會是一件好事嗎?

  • they'll be able to make better decisions in the real world than we can.

    我當然希望如此。

  • So is that a good thing?

    人類的全部文明, 我們所珍視的一切,

  • Well, I hope so.

    都是基於我們的智慧。

  • Our entire civilization, everything that we value,

    如果我們能獲得更強大的智慧,

  • is based on our intelligence.

    那人類將無所不能了。

  • And if we had access to a lot more intelligence,

    我在想,到時後就像 一些人所描述的那樣,

  • then there's really no limit to what the human race can do.

    這會是人類歷史上最重要的事件。

  • And I think this could be, as some people have described it,

    那為什麽有的人會說出 以下的言論呢?

  • the biggest event in human history.

    說人工智慧將是人類的末日呢?

  • So why are people saying things like this,

    這是新鮮事嗎?

  • that AI might spell the end of the human race?

    這僅僅只是伊隆馬斯克、比爾蓋茲、 史蒂芬霍金的新發明嗎?

  • Is this a new thing?

    實際上不是,這個概念 已經存在很長的時間了。

  • Is it just Elon Musk and Bill Gates and Stephen Hawking?

    請看這段話:

  • Actually, no. This idea has been around for a while.

    「即便我們能讓機器屈從於我們,

  • Here's a quotation:

    比如說,在重要時刻關掉它。」

  • "Even if we could keep the machines in a subservient position,

    我等會兒會再來討論 「關機」這一話題。

  • for instance, by turning off the power at strategic moments" --

    「我們作為人類,仍應懷着謙卑......」

  • and I'll come back to that "turning off the power" idea later on --

    這段話是誰說的呢? 是艾倫 · 圖靈在 1951 年說的。

  • "we should, as a species, feel greatly humbled."

    眾所皆知艾倫 · 圖靈是計算機科學之父,

  • So who said this? This is Alan Turing in 1951.

    並且從很多方面來講, 他也是人工智慧之父。

  • Alan Turing, as you know, is the father of computer science

    所以,當我們在思考「創造出 比自己更聰明的物種」這個問題時,

  • and in many ways, the father of AI as well.

    我們不妨將它稱為「大猩猩問題」。

  • So if we think about this problem,

    因為大猩猩的祖先們 在幾百萬年前就親歷此境,

  • the problem of creating something more intelligent than your own species,

    我們可以去問大猩猩們:

  • we might call this "the gorilla problem,"

    「這是不是一個好主意?」

  • because gorillas' ancestors did this a few million years ago,

    圖片中,牠們正在開會討論 那麽做是不是一個好主意,

  • and now we can ask the gorillas:

    過了一會兒,牠們總結出:「不。」

  • Was this a good idea?

    這是個很爛的主意──

  • So here they are having a meeting to discuss whether it was a good idea,

    作為靈長類的我們正岌岌可危。

  • and after a little while, they conclude, no,

    你可以從牠們的眼神中 看到存亡攸關的憂傷。

  • this was a terrible idea.

    (笑聲)

  • Our species is in dire straits.

    「創造出比你自己更聰明的物種 並不是什麽妙計」

  • In fact, you can see the existential sadness in their eyes.

    這種感覺很倒胃口。

  • (Laughter)

    那我們能做些什麽呢?

  • So this queasy feeling that making something smarter than your own species

    其實,除非停止人工智慧的研究, 否則束手無策。

  • is maybe not a good idea --

    因為我所提到的人工智慧的各種裨益,

  • what can we do about that?

    也因為我是人工智慧的研究人員,

  • Well, really nothing, except stop doing AI,

    我可不同意就此止步。

  • and because of all the benefits that I mentioned

    實際上,我想一直研究人工智慧。

  • and because I'm an AI researcher,

    所以我們需要更加明確問題所在。

  • I'm not having that.

    這個問題到底是什麽呢?

  • I actually want to be able to keep doing AI.

    為什麽更強大的人工智慧 可能會是個災難呢?

  • So we actually need to nail down the problem a bit more.

    還有一句名言:

  • What exactly is the problem?

    「我們最好確保我們向機器發出的指令 與我們的真正目的相吻合。」

  • Why is better AI possibly a catastrophe?

    這句話是諾伯特 · 維納在 1960 年說的,

  • So here's another quotation:

    就在他看完一個早期的學習系統 (Learning System)之後。

  • "We had better be quite sure that the purpose put into the machine

    這個系統在學習如何能把 西洋棋下得比發明它的人更好。

  • is the purpose which we really desire."

    但如出一轍的一句話,

  • This was said by Norbert Wiener in 1960,

    邁達斯國王也說過。

  • shortly after he watched one of the very early learning systems

    他說:「我希望我觸碰的 所有東西都變成金子。」

  • learn to play checkers better than its creator.

    結果他真的獲得了點石成金的能力。

  • But this could equally have been said

    可以說,這就是他給機器下的指令。

  • by King Midas.

    結果他的食物、飲料 和家人都變成了金子,

  • King Midas said, "I want everything I touch to turn to gold,"

    最後他死於痛苦與饑餓當中。

  • and he got exactly what he asked for.

    所以我們把這類問題叫做 「邁達斯國王問題」,

  • That was the purpose that he put into the machine,

    這個比喻是要說明這種 不符合實際需求的 「目的」。

  • so to speak,

    用現代的術語來說,我們把它稱為 「價值取向不一致問題」。

  • and then his food and his drink and his relatives turned to gold

    「設錯了目標」不是唯一的問題,

  • and he died in misery and starvation.

    還有其他的。

  • So we'll call this "the King Midas problem"

    如果你給機器人設了個目標,

  • of stating an objective which is not, in fact,

    即使簡單如「去把咖啡端來。」

  • truly aligned with what we want.

    那機器人會對自己說:

  • In modern terms, we call this "the value alignment problem."

    「什麼會讓我無法去拿咖啡?

  • Putting in the wrong objective is not the only part of the problem.

    說不定有人會把我關機;

  • There's another part.

    好,那我要想辦法阻止,

  • If you put an objective into a machine,

    我得讓我的「關機」開關失效。

  • even something as simple as, "Fetch the coffee,"

    我得盡一切可能防衛自己,

  • the machine says to itself,

    免得別人干涉我去達成 所被賦予的任務。」

  • "Well, how might I fail to fetch the coffee?

    這種專注的行事,以一種 極端自我保護的模式在執行,

  • Someone might switch me off.

    實際上與我們人類 想要的目標並不一致。

  • OK, I have to take steps to prevent that.

    這就是我們面臨的問題。

  • I will disable my 'off' switch.

    而這就是這場演講的 核心想法,也是價值所在。

  • I will do anything to defend myself against interference

    如果你想從這場演講中汲取什麽,

  • with this objective that I have been given."

    那你只要記得: 如果死了,就不能端咖啡了。

  • So this single-minded pursuit

    (笑聲)

  • in a very defensive mode of an objective that is, in fact,

    這很簡單,記住就行了, 每天早晚覆誦三遍。

  • not aligned with the true objectives of the human race --

    (笑聲)

  • that's the problem that we face.

    實際上,這正是電影 《2001太空漫步》的劇情。

  • And in fact, that's the high-value takeaway from this talk.

    HAL 有一個目標,一個任務,

  • If you want to remember one thing,

    但這個目標與人類的目標不一致,

  • it's that you can't fetch the coffee if you're dead.

    最後導致了衝突。

  • (Laughter)

    幸運的是, HAL 並沒有超級智慧,

  • It's very simple. Just remember that. Repeat it to yourself three times a day.

    它挺聰明的, 但還是比不過人類戴夫,

  • (Laughter)

    戴夫可以把 HAL 關掉。

  • And in fact, this is exactly the plot

    但我們可能就沒有這麽幸運了。

  • of "2001: [A Space Odyssey]"

    那我們應該怎麽辦呢?

  • HAL has an objective, a mission,

    我想要重新定義人工智慧,

  • which is not aligned with the objectives of the humans,

    不再囿於傳統的概念:

  • and that leads to this conflict.

    能明智地達成目標的機器。

  • Now fortunately, HAL is not superintelligent.

    新的定義涉及三條原則。

  • He's pretty smart, but eventually Dave outwits him

    第一個原則是利他主義原則,

  • and manages to switch him off.

    也就是說,機器的唯一目標

  • But we might not be so lucky.

    就是要最大化地實現 人類的目標、人類的價值。

  • So what are we going to do?

    這種價值不是指多愁善感 或者假裝乖巧,

  • I'm trying to redefine AI

    而是指人類所嚮往、追求的生活, 無論現狀如何。

  • to get away from this classical notion

    事實上,這樣就違反了艾西莫夫定律,

  • of machines that intelligently pursue objectives.

    定律裡的機器人必須維護自己的生存。

  • There are three principles involved.

    而在這條原則裡 機器對自身生存與否毫不關心。

  • The first one is a principle of altruism, if you like,

    第二個原則,不妨稱之為謙遜原則。

  • that the robot's only objective

    這一條對製造出安全的機器人十分重要。

  • is to maximize the realization of human objectives,

    它是指機器人不知道人類的價值是什麽,

  • of human values.

    它只知道將該價值最大化, 但卻不知道該價值究竟是什麽。

  • And by values here I don't mean touchy-feely, goody-goody values.

    這就避免了「追求單一目的 而不知變通」的現象。

  • I just mean whatever it is that the human would prefer

    這種不確定性就變得很重要了。

  • their life to be like.

    為了對我們有益,

  • And so this actually violates Asimov's law

    機械就得大概明白我們想要什麽。

  • that the robot has to protect its own existence.

    它要獲取這類信息,主要是 透過觀察人類的決策,

  • It has no interest in preserving its existence whatsoever.

    所以我們的決策會揭露 我們生活的意願,

  • The second law is a law of humility, if you like.

    所以,這三條原則,

  • And this turns out to be really important to make robots safe.

    讓我們來看看要如何 應用到圖靈所說的問題:

  • It says that the robot does not know

    「你能不能將機器關掉?」

  • what those human values are,

    這是 PR2 機器人,

  • so it has to maximize them, but it doesn't know what they are.

    這是我們實驗室裡的其中一台,

  • And that avoids this problem of single-minded pursuit

    它的背面有一個大大的紅色開關。

  • of an objective.

    那問題來了:它會讓你把它關掉嗎?

  • This uncertainty turns out to be crucial.

    如果我們用傳統的定義製造它,

  • Now, in order to be useful to us,

    我們給它一個「去拿咖啡」的目標, 它會想:「我必須去拿咖啡,

  • it has to have some idea of what we want.

    但如果我死了,就不能拿咖啡了。」

  • It obtains that information primarily by observation of human choices,

    看來, PR2 聽過我的演講了,

  • so our own choices reveal information

    因此它說:「我必須讓自己的開關失靈,

  • about what it is that we prefer our lives to be like.

    可能還要通過電擊把那些在 星巴克裡干擾我的人都擊暈。」

  • So those are the three principles.

    (笑聲)

  • Let's see how that applies to this question of:

    這無法避免,對吧?

  • "Can you switch the machine off?" as Turing suggested.

    這種失敗看起來是必然的,

  • So here's a PR2 robot.

    因為機器人會遵循一個 十分明確的目標。

  • This is one that we have in our lab,

    那如果機器對目標 不那麽確定會發生什麽呢?

  • and it has a big red "off" switch right on the back.

    那它的思路就不一樣了。

  • The question is: Is it going to let you switch it off?

    它會說:「好的,人類可能會把我關掉,

  • If we do it the classical way,

    但只有我做錯事了,才會把我關掉。

  • we give it the objective of, "Fetch the coffee, I must fetch the coffee,

    沒錯,我真的不知道什麽才是錯,

  • I can't fetch the coffee if I'm dead,"

    但我知道我不該做錯的事。」

  • so obviously the PR2 has been listening to my talk,

    這就是第一和第二原則。

  • and so it says, therefore, "I must disable my 'off' switch,

    「所以我應該讓人類把我關掉。」

  • and probably taser all the other people in Starbucks

    事實上你可以推斷出機器人為了 允許讓人類關掉它所包含的動機,

  • who might interfere with me."

    而且這與根本目標的 不確定性程度直接相關。

  • (Laughter)

    當機器被關閉後,

  • So this seems to be inevitable, right?

    第三條原則就起作用了。

  • This kind of failure mode seems to be inevitable,

    機器開始學習它應追求的目標,

  • and it follows from having a concrete, definite objective.

    因為它知道它剛才做的事是不對的。

  • So what happens if the machine is uncertain about the objective?

    實際上,我們可以適當地 使用些希臘字母,

  • Well, it reasons in a different way.

    就像數學家們經常做的那樣,

  • It says, "OK, the human might switch me off,

    直接證明這一個理論:這樣的 機器人對人類是絕對有利的。

  • but only if I'm doing something wrong.

    可以證明如此設計出來的機器人, 對我們的生活是是有益的。

  • Well, I don't really know what wrong is,

    這個例子很簡單,

  • but I know that I don't want to do it."

    但它是我們嘗試實現 能與人類和諧共處的 AI 的第一步。

  • So that's the first and second principles right there.

    現在來看第三個原則,

  • "So I should let the human switch me off."

    我知道各位可能還在為 這一個原則傷腦筋。

  • And in fact you can calculate the incentive that the robot has

    你可能會想:「你懂的, 我行為舉止比較差勁。

  • to allow the human to switch it off,

    我的機器人可不能被我帶壞。

  • and it's directly tied to the degree

    我有時後會大半夜偷偷摸摸地 從冰箱裡找東西吃,

  • of uncertainty about the underlying objective.

    東瞅瞅,西摸摸。」

  • And then when the machine is switched off,

    有各種各樣的事 你是不希望機器人去做的。

  • that third principle comes into play.

    但實際上不是那樣。

  • It learns something about the objectives it should be pursuing,

    你行為不檢,

  • because it learns that what it did wasn't right.

    不代表機器人就得有樣學樣。

  • In fact, we can, with suitable use of Greek symbols,

    它會去嘗試理解你做事的動機,

  • as mathematicians usually do,

    而且可能會在合適的情況下 幫助你、制止你。

  • we can actually prove a theorem

    但這仍然十分困難。

  • that says that such a robot is provably beneficial to the human.

    實際上,我們是要讓機器

  • You are provably better off with a machine that's designed in this way

    為任何人、任何一種 可能的生活去預測:

  • than without it.

    他們更想怎樣?更想要什麽?

  • So this is a very simple example, but this is the first step

    這涉及到諸多困難,

  • in what we're trying to do with human-compatible AI.

    我不認為這會很快地就被解決。

  • Now, this third principle,

    實際上,真正的困難是我們自己。

  • I think is the one that you're probably scratching your head over.

    就像我剛說的那樣, 我們做事不守規矩。

  • You're probably thinking, "Well, you know, I behave badly.

    我們當中就有人是非常惡劣的。

  • I don't want my robot to behave like me.

    如前所說,機器人 未必得要複製那些行為。

  • I sneak down in the middle of the night and take stuff from the fridge.

    機器人沒有自己的目標,

  • I do this and that."

    它是完全利他的。

  • There's all kinds of things you don't want the robot doing.

    它的誕生不僅僅是為了去滿足 某一個人、某一個用戶的欲望,

  • But in fact, it doesn't quite work that way.

    而是去尊重所有人的意願。

  • Just because you behave badly

    所以它懂得抵制一些惡劣的行為,

  • doesn't mean the robot is going to copy your behavior.

    它甚至能理解你為什麼惡劣,比如說,

  • It's going to understand your motivations and maybe help you resist them,

    如果你是一個邊境護照官員, 你可能會收取賄賂,

  • if appropriate.

    因為你得養家、供孩子們上學。

  • But it's still difficult.

    機器人能理解這一點, 但不代表它也會學你偷錢,

  • What we're trying to do, in fact,

    它反而會幫助你去供孩子們上學。

  • is to allow machines to predict for any person and for any possible life

    我們的計算能力也是有限的。

  • that they could live,

    李世石是一個傑出的圍棋大師,

  • and the lives of everybody else:

    但他還是輸了。

  • Which would they prefer?

    如果我們仔細觀察他的棋路, 他下錯了那幾步以致輸棋,

  • And there are many, many difficulties involved in doing this;

    但這不意味著他想要輸。

  • I don't expect that this is going to get solved very quickly.

    所以要理解他的行為,

  • The real difficulties, in fact, are us.

    我們得從人類認知的模型回推過來,

  • As I have already mentioned, we behave badly.

    它包含了我們計算能力上的局限,

  • In fact, some of us are downright nasty.

    是一個很覆雜的模型。

  • Now the robot, as I said, doesn't have to copy the behavior.

    但我們仍然可以嘗試去理解。

  • The robot does not have any objective of its own.

    可能對於我這樣的 AI 研究人員來說,

  • It's purely altruistic.

    最大的困難是,人有很多種,

  • And it's not designed just to satisfy the desires of one person, the user,

    所以機器必須想辦法去協調、 權衡不同人之間的喜好、需求,

  • but in fact it has to respect the preferences of everybody.

    而要做到這一點有多種不同的方法。

  • So it can deal with a certain amount of nastiness,

    經濟學家、社會學家、 道德哲學家都理解這一點,

  • and it can even understand that your nastiness, for example,

    我們正積極地尋求合作。

  • you may take bribes as a passport official

    讓我們來看看,如果我們把這一步 走錯了會怎麽樣。

  • because you need to feed your family and send your kids to school.

    比如說,你可能會與你的 人工智慧助理有這樣的對話,

  • It can understand that; it doesn't mean it's going to steal.

    這樣的人工智慧可能幾年內就會出現。

  • In fact, it'll just help you send your kids to school.

    可以把它想成是強化版的 Siri 。

  • We are also computationally limited.

    Siri 對你說:「你老婆打電話 提醒你別忘了今天的晚宴。」

  • Lee Sedol is a brilliant Go player,

    當然你早就忘了這回事:

  • but he still lost.

    「什麽?什麽晚宴?你在說什麽?」

  • So if we look at his actions, he took an action that lost the game.

    「呃.....今晚 7 點 慶祝結婚 20 周年。」

  • That doesn't mean he wanted to lose.

    「我可去不了, 我晚上 7 點半要見秘書長。

  • So to understand his behavior,

    怎麽會這樣呢?」

  • we actually have to invert through a model of human cognition

    「呃,我可是提醒過你的, 但你沒有理會我的建議。」

  • that includes our computational limitations -- a very complicated model.

    「我該怎麽辦呢?我可不能跟秘書長說 我有事,沒空見他。」

  • But it's still something that we can work on understanding.

    「別擔心。我已經安排了, 讓他的航班延誤。」

  • Probably the most difficult part, from my point of view as an AI researcher,

    (笑聲)

  • is the fact that there are lots of us,

    「用某種電腦故障。」

  • and so the machine has to somehow trade off, weigh up the preferences

    (笑聲)

  • of many different people,

    「真的嗎?這個你也能做到?」

  • and there are different ways to do that.

    「秘書長很不好意思,跟你道歉,

  • Economists, sociologists, moral philosophers have understood that,

    並邀請你明天中午吃飯。」

  • and we are actively looking for collaboration.

    (笑聲)

  • Let's have a look and see what happens when you get that wrong.

    所以這裡談的價值觀就有點問題了,

  • So you can have a conversation, for example,

    這顯然是在遵循我老婆的價值觀,

  • with your intelligent personal assistant

    也就是「老婆開心,生活舒心」。

  • that might be available in a few years' time.

    (笑聲)

  • Think of a Siri on steroids.

    它也有可能發展成另一種情況。

  • So Siri says, "Your wife called to remind you about dinner tonight."

    你忙碌一天,回到家裏,

  • And of course, you've forgotten. "What? What dinner?

    電腦對你說:「今天很忙喔?」

  • What are you talking about?"

    「是啊,我連午飯都沒來得及吃。」

  • "Uh, your 20th anniversary at 7pm."

    「那你一定很餓了吧。」

  • "I can't do that. I'm meeting with the secretary-general at 7:30.

    「快餓暈了。你能做點晚飯嗎?」

  • How could this have happened?"

    「有一件事我得告訴你。」

  • "Well, I did warn you, but you overrode my recommendation."

    (笑聲)

  • "Well, what am I going to do? I can't just tell him I'm too busy."

    「南蘇丹人民的情況 比你更緊急,更需要照顧。」

  • "Don't worry. I arranged for his plane to be delayed."

    (笑聲)

  • (Laughter)

    「所以我要走了。你自己做飯去吧。」

  • "Some kind of computer malfunction."

    (笑聲)

  • (Laughter)

    我們得解決這類的問題,

  • "Really? You can do that?"

    我也很期待能解決這樣的問題。

  • "He sends his profound apologies

    我們有理由感到樂觀。

  • and looks forward to meeting you for lunch tomorrow."

    理由之一是,

  • (Laughter)

    我們有大量的數據資料。

  • So the values here -- there's a slight mistake going on.

    記住,我說過機器將能夠 閱讀所有人類寫下來的東西。

  • This is clearly following my wife's values

    而我們寫下的文字大都類似於

  • which is "Happy wife, happy life."

    「人類做了一些事情 導致其他人對此感到沮喪」。

  • (Laughter)

    所以機器可以從 大量的數據中去學習。

  • It could go the other way.

    同時從經濟的角度, 我們也有足夠的動機去做好這件事。

  • You could come home after a hard day's work,

    想像一下,你家裡有個居家機器人。

  • and the computer says, "Long day?"

    而你又得加班, 機器人得給孩子們做飯,

  • "Yes, I didn't even have time for lunch."

    孩子們很餓, 但冰箱裡什麽都沒有。

  • "You must be very hungry."

    然後機器人看到了家裡的貓。

  • "Starving, yeah. Could you make some dinner?"

    (笑聲)

  • "There's something I need to tell you."

    機器人還沒學透人類的價值觀。

  • (Laughter)

    所以它不知道, 貓的情感價值大於其營養價值。

  • "There are humans in South Sudan who are in more urgent need than you."

    (笑聲)

  • (Laughter)

    接下來會發生什麽事?

  • "So I'm leaving. Make your own dinner."

    頭版頭條可能會是這樣:

  • (Laughter)

    「瘋狂機器人煮了貓咪當晚餐!」

  • So we have to solve these problems,

    這場意外就足以結束 整個居家機器人的產業。

  • and I'm looking forward to working on them.

    所以在我們實現超級 AI 之前, 我們有足夠的動機把它做對做好。

  • There are reasons for optimism.

    總結來說:

  • One reason is,

    我事實上想要改變人工智慧的定義,

  • there is a massive amount of data.

    這樣我們就可以製造出 對我們有益無害的機器人。

  • Because remember -- I said they're going to read everything

    這三個原則是:

  • the human race has ever written.

    機器是利他的,

  • Most of what we write about is human beings doing things

    只想著實現我們的目標,

  • and other people getting upset about it.

    但它不確定我們的目標是什麽,

  • So there's a massive amount of data to learn from.

    並且它會觀察我們,

  • There's also a very strong economic incentive

    從中學習我們想要的究竟是什麽。

  • to get this right.

    希望在這個過程中, 我們也能學會成為更好的人。

  • So imagine your domestic robot's at home.

    謝謝大家。

  • You're late from work again and the robot has to feed the kids,

    (掌聲)

  • and the kids are hungry and there's nothing in the fridge.

    克里斯安德森:非常有意思,斯圖爾特。

  • And the robot sees the cat.

    趁工作人員為下一位講者佈置的時候,

  • (Laughter)

    我們先站在這裡聊幾句。

  • And the robot hasn't quite learned the human value function properly,

    我有幾個問題。

  • so it doesn't understand

    將「無知」編寫到程式中, 這種思想真的很有衝擊力。

  • the sentimental value of the cat outweighs the nutritional value of the cat.

    當機器人有超級智慧時,

  • (Laughter)

    還有什麽東西能阻檔機器人閱讀書籍,

  • So then what happens?

    並了解到:博學比無知要好得多,

  • Well, it happens like this:

    進而改變它的目標, 重新編寫自己的程式呢?

  • "Deranged robot cooks kitty for family dinner."

    斯圖爾特拉塞爾:是的, 我們想要它去學習,就像我說的,

  • That one incident would be the end of the domestic robot industry.

    讓機器人學習我們的目標,

  • So there's a huge incentive to get this right

    只有在理解得越正確的時候, 它們才會更明確我們要的東西,

  • long before we reach superintelligent machines.

    佐證擺在那裡,

  • So to summarize:

    並且我們使它能夠正確解讀這些目標。

  • I'm actually trying to change the definition of AI

    比如說,它能夠從書中的佐證 判斷出那些富含偏見的書,

  • so that we have provably beneficial machines.

    像是只講述國王、王子, 和男性精英白人之類的書。

  • And the principles are:

    所以這是一個複雜的問題,

  • machines that are altruistic,

    但當它更深入地學習我們的目標時,

  • that want to achieve only our objectives,

    它會變得越來越有用。

  • but that are uncertain about what those objectives are,

    CA:所以它十分複雜, 遠不足以濃縮成一條法則嗎?

  • and will watch all of us

    像是,把這樣的命令燒録進去:

  • to learn more about what it is that we really want.

    「如果人類想把我關掉,

  • And hopefully in the process, we will learn to be better people.

    我要服從。我要服從。」

  • Thank you very much.

    SR:絕對不行。

  • (Applause)

    那將是一個很糟糕的主意。

  • Chris Anderson: So interesting, Stuart.

    試想一下,你有一輛無人駕駛汽車,

  • We're going to stand here a bit because I think they're setting up

    你想讓它送你五歲的孩子去幼稚園。

  • for our next speaker.

    你會希望你五歲的孩子 在汽車運行的過程中將它關閉嗎?

  • A couple of questions.

    應該不會吧。

  • So the idea of programming in ignorance seems intuitively really powerful.

    所以它得理解

  • As you get to superintelligence,

    下指令的人有多理智、有多講道理。

  • what's going to stop a robot

    這個人越理智,

  • reading literature and discovering this idea that knowledge

    它就越願意被你關掉。

  • is actually better than ignorance

    如果這個人是完全思緒混亂 或者甚至是有惡意的,

  • and still just shifting its own goals and rewriting that programming?

    那它就不太願意被你關掉了。

  • Stuart Russell: Yes, so we want it to learn more, as I said,

    CA:好吧。斯圖爾特,我得說,

  • about our objectives.

    我真的希望你為我們所有人, 找到解決的辦法。

  • It'll only become more certain as it becomes more correct,

    很感謝你的演講。 十分精彩。

  • so the evidence is there

    SR:謝謝。

  • and it's going to be designed to interpret it correctly.

    CA:謝謝。

  • It will understand, for example, that books are very biased

    (掌聲)

  • in the evidence they contain.

  • They only talk about kings and princes

  • and elite white male people doing stuff.

  • So it's a complicated problem,

  • but as it learns more about our objectives

  • it will become more and more useful to us.

  • CA: And you couldn't just boil it down to one law,

  • you know, hardwired in:

  • "if any human ever tries to switch me off,

  • I comply. I comply."

  • SR: Absolutely not.

  • That would be a terrible idea.

  • So imagine that you have a self-driving car

  • and you want to send your five-year-old

  • off to preschool.

  • Do you want your five-year-old to be able to switch off the car

  • while it's driving along?

  • Probably not.

  • So it needs to understand how rational and sensible the person is.

  • The more rational the person,

  • the more willing you are to be switched off.

  • If the person is completely random or even malicious,

  • then you're less willing to be switched off.

  • CA: All right. Stuart, can I just say,

  • I really, really hope you figure this out for us.

  • Thank you so much for that talk. That was amazing.

  • SR: Thank you.

  • (Applause)

This is Lee Sedol.

譯者: 易帆 余 審譯者: Wilde Luo

Subtitles and vocabulary

Click the word to look it up Click the word to find further inforamtion about it

A2 US TED 機器人 什麽 人類 目標 關掉

【TED】斯圖亞特-羅素:創造更安全的AI的3個原則(創造更安全的AI的3個原則|斯圖亞特-羅素)。 (【TED】Stuart Russell: 3 principles for creating safer AI (3 principles for creating safer AI | Stuart Russell))

  • 68 14
    Zenn posted on 2021/01/14
Video vocabulary