-
我與一群數學家、哲學家及電腦科學家共事,我們坐在一起思考機器智慧的未來以及其他問題。有人認為這屬於科幻小說領域,既不切實際又瘋狂。但我想說的是-好,我們看看現代人類的情況(笑聲)。這是現代人類的常態,但如果我們思考這一點,事實上人類只是抵達地球不久的訪客。假設地球在一年前誕生,那麼人類的存在相當於10分鐘,工業世代在2秒前開始。另一個探討方式是,思考近一萬年來的全球GDP(國內生產毛額),我費了點功夫繪製了這張圖表,看起來像這樣(笑聲)。這對正常狀態來說是相當奇特的形狀,我肯定不想坐在上面(笑聲)。我們思考一下:造成這種異常的原因為何?有些人會說是科技。確實如此,科技隨著人類歷史而發展,現在科技正以飛快的速度進步,這是近因,這也是為何目前的生產力相當高,但我想進一步探索根本原因。
展開英文
收合英文
-
I work with a bunch of mathematicians, philosophers and computer scientists, and we sit around and think about the future of machine intelligence, among other things. Some people think that some of these things are sort of science fiction-y, far out there, crazy. But I like to say, okay, let's look at the modern human condition. (Laughter) This is the normal way for things to be. But if we think about it, we are actually recently arrived guests on this planet, the human species. Think about if Earth was created one year ago, the human species, then, would be 10 minutes old. The industrial era started two seconds ago. Another way to look at this is to think of world GDP over the last 10,000 years, I've actually taken the trouble to plot this for you in a graph. It looks like this. (Laughter) It's a curious shape for a normal condition. I sure wouldn't want to sit on it. (Laughter) Let's ask ourselves, what is the cause of this current anomaly? Some people would say it's technology. Now it's true, technology has accumulated through human history, and right now, technology advances extremely rapidly -- that is the proximate cause, that's why we are currently so very productive. But I like to think back further to the ultimate cause.
-
看看這兩位相當傑出的紳士:這位是Kanzi,牠能掌握200個詞彙,相當驚人的壯舉。以及愛德華.維騰,他引起第二次超弦革命。如果我們進行深入探索,這是我們的發現:基本上是相同的東西,一個稍微大一點,或許其中存在一些特殊的連結方式,但這些無形的差異不會太複雜,因為從我們最後一位共同祖先算起只經歷了25萬代。我們知道複雜的機制需要長時間演化,因此一些相對微小的變化使我們從Kanzi變成維騰,從以掉落的樹枝當武器變成發射洲際彈道式導彈。因此顯然我們所達成的一切,以及所有我們關心的事物都取決於人腦中相對微小的改變。理所當然的推論是:任何進一步的改變、能造成思想基體的顯著改變都可能產生巨大影響。我的一些同事認為我們即將發現某種能深刻改變思想基體的東西,那就是機器的超級智慧。人工智慧通常是指將指令輸入一個箱子,你需要程式設計師費心將知識轉變成指令。你建立這些專門系統,它們可用於某些目的,但這些系統相當生硬,你無法延伸它的規模,基本上你僅能獲得事先輸入的資訊。但從那時起,人工智慧領域發生革命性改變,如今的目標在於機器學習。因此我們並非研究知識的展現,我們撰寫具有學習能力的演算法,通常來自原始的感知數據,基本上如同嬰兒的學習方式,結果是產生不侷限於某個領域的人工智慧。同一個系統可以學習在任何兩種語言之間翻譯,或學習玩Atari遊戲機的任何一款遊戲。
展開英文
收合英文
-
Look at these two highly distinguished gentlemen: We have Kanzi -- he's mastered 200 lexical tokens, an incredible feat. And Ed Witten unleashed the second superstring revolution. If we look under the hood, this is what we find: basically the same thing. One is a little larger, it maybe also has a few tricks in the exact way it's wired. These invisible differences cannot be too complicated, however, because there have only been 250,000 generations since our last common ancestor. We know that complicated mechanisms take a long time to evolve. So a bunch of relatively minor changes take us from Kanzi to Witten, from broken-off tree branches to intercontinental ballistic missiles. So this then seems pretty obvious that everything we've achieved, and everything we care about, depends crucially on some relatively minor changes that made the human mind. And the corollary, of course, is that any further changes that could significantly change the substrate of thinking could have potentially enormous consequences. Some of my colleagues think we're on the verge of something that could cause a profound change in that substrate, and that is machine superintelligence. Artificial intelligence used to be about putting commands in a box. You would have human programmers that would painstakingly handcraft knowledge items. You build up these expert systems, and they were kind of useful for some purposes, but they were very brittle, you couldn't scale them. Basically, you got out only what you put in. But since then, a paradigm shift has taken place in the field of artificial intelligence. Today, the action is really around machine learning. So rather than handcrafting knowledge representations and features, we create algorithms that learn, often from raw perceptual data. Basically the same thing that the human infant does. The result is A.I. that is not limited to one domain -- the same system can learn to translate between any pairs of languages, or learn to play any computer game on the Atari console.
-
現在,當然人工智慧仍無法像人類一樣擁有強大的跨領域學習與計畫能力,大腦皮層仍擁有一些我們還不知道如何複製在機器上的演算技巧。因此問題是:還有多久我們才能複製這些技巧?幾年前我們對全球頂尖的人工智慧專家做了一項調查,試圖瞭解他們的想法,其中一個問題是:「你認為哪一年我們將擁有50%的機率使機器智慧達到人類級?」我們把「人類級」定義為有能力將任何任務以成年人的水準執行,因此真正的「人類級」不僅限於某些特定領域,答案的中間值是2040或2050年,取決於我們詢問的專家屬於什麼群體。也許過了很久才會實現,也許很快就會實現,沒有人能知道確切時間。我們知道的是機器基體處理資訊的最終極限比生物組織的極限大得多,這與物理原理有關。一個生物神經元發出的脈衝頻率大約200赫茲,相當於每秒200次,但即使是目前的電晶體,脈衝頻率都以千兆赫計算。神經元在軸突中的傳輸速度緩慢,頂多每秒100公尺,但在電腦中訊號以光速傳播。還有尺寸限制,就像人類大腦必須能裝進顱骨內,但電腦可以跟倉庫一樣大,或更大。因此超級智慧的潛能正處於潛伏狀態,就像原子能潛伏在人類歷史中耐心等到1945年被發現。在這個世紀中,科學家或許能喚醒人工智慧的力量,我認為我們或許有機會見證人工智慧的爆發。
展開英文
收合英文
-
Now of course, A.I. is still nowhere near having the same powerful, cross-domain ability to learn and plan as a human being has. The cortex still has some algorithmic tricks that we don't yet know how to match in machines. So the question is, how far are we from being able to match those tricks? A couple of years ago, we did a survey of some of the world's leading A.I. experts, to see what they think, and one of the questions we asked was, "By which year do you think there is a 50 percent probability that we will have achieved human-level machine intelligence?" We defined human-level here as the ability to perform almost any job at least as well as an adult human, so real human-level, not just within some limited domain. And the median answer was 2040 or 2050,depending on precisely which group of experts we asked. Now, it could happen much, much later, or sooner, the truth is nobody really knows. What we do know is that the ultimate limit to information processing in a machine substrate lies far outside the limits in biological tissue. This comes down to physics. A biological neuron fires, maybe, at 200 hertz, 200 times a second. But even a present-day transistor operates at the Gigahertz. Neurons propagate slowly in axons, 100 meters per second, tops. But in computers, signals can travel at the speed of light. There are also size limitations, like a human brain has to fit inside a cranium, but a computer can be the size of a warehouse or larger. So the potential for superintelligence lies dormant in matter, much like the power of the atom lay dormant throughout human history, patiently waiting there until 1945. In this century, scientists may learn to awaken the power of artificial intelligence. And I think we might then see an intelligence explosion.
-
當大多數人思考何謂聰明、何謂愚蠢時,我腦海裡浮現一個像這樣的畫面:一邊是一個鄉下傻子,在遙遠的另一邊是愛德華.維騰或亞伯特.愛因斯坦,或任何你喜愛的大師。但我認為以人工智慧的觀點來看,真正的畫面或許比較像這樣:人工智慧從這一點開始,零智慧;經多年辛苦研究之後,也許最後我們能達到老鼠級的人工智慧,可在雜亂的環境中找到路,就像老鼠一樣。再經多年辛苦研究及許多投資之後,也許最後我們能達到黑猩猩級的人工智慧。再經更多年辛苦研究之後,我們達到鄉下傻子級的人工智慧。再過不久,我們將超越愛德華.維騰。這列火車不會停在人類村這一站,它很可能呼嘯而過。這具有深遠意義,尤其是涉及權力問題時。例如,黑猩猩很強壯,若是公平比較,黑猩猩大約比健康男性強壯兩倍。然而,Kanzi和同伴的命運多半取決於人類的作為,而非黑猩猩本身的作為。一旦超級智慧出現,人類的命運或許將取決於超級智慧的作為。思考一下:機器智慧將是人類所需的最後一項發明,之後機器將比人類更擅長發明,它們將以「數位時間尺度」進行,基本上這意味著縮短未來的距離。思考你曾經想像過的所有瘋狂科技,也許人類將擁有充分時間發展出以下事物:防止衰老、太空殖民、自行複製的奈米機器人,或將思想上傳電腦。這都是科幻小說的內容,但仍符合物理法則。
展開英文
收合英文
-
Now most people, when they think about what is smart and what is dumb, I think have in mind a picture roughly like this. So at one end we have the village idiot, and then far over at the other side we have Ed Witten, or Albert Einstein, or whoever your favorite guru is. But I think that from the point of view of artificial intelligence, the true picture is actually probably more like this: AI starts out at this point here, at zero intelligence, and then, after many, many years of really hard work, maybe eventually we get to mouse-level artificial intelligence, something that can navigate cluttered environments as well as a mouse can. And then, after many, many more years of really hard work, lots of investment, maybe eventually we get to chimpanzee-level artificial intelligence. And then, after even more years of really, really hard work, we get to village idiot artificial intelligence. And a few moments later, we are beyond Ed Witten. The train doesn't stop at Humanville Station. It's likely, rather, to swoosh right by. Now this has profound implications, particularly when it comes to questions of power. For example, chimpanzees are strong -- pound for pound, a chimpanzee is about twice as strong as a fit human male. And yet, the fate of Kanzi and his pals depends a lot more on what we humans do than on what the chimpanzees do themselves. Once there is superintelligence, the fate of humanity may depend on what the superintelligence does. Think about it: Machine intelligence is the last invention that humanity will ever need to make. Machines will then be better at inventing than we are, and they'll be doing so on digital timescales. What this means is basically a telescoping of the future. Think of all the crazy technologies that you could have imagined maybe humans could have developed in the fullness of time: cures for aging, space colonization, self-replicating nanobots or uploading of minds into computers, all kinds of science fiction-y stuff that's nevertheless consistent with the laws of physics.
-
超級智慧能發展出這一切,也許相當迅速。這種技術成熟的超級智慧將會非常強大,至少在某些處境中能得到它想要的東西,我們將擁有以超級智慧的偏好所塑造的未來。一個很好的問題是:這些偏好是什麼?這讓情況變得更棘手。在這個領域裡前進,我們首先得避免機器智慧人格化。這有點諷刺,因為每一篇關於未來人工智慧的報導都會附上這張照片,因此我認為我們必須以更抽象的方式考慮這個問題,而非以好萊塢電影中的場景考慮。我們需要將智慧視為優化過程,一個將未來引導至特定組態的過程。超級智慧是相當強大的優化過程,它擅於利用現有工具達成目標狀態,這意味著在高智慧及擁有對人類來說有價值或意義的目標之間沒有必然的聯繫。假設我們給予人工智慧的目標是讓人類發笑,當人工智慧尚未成熟時,它會做出有用或好笑的動作引使用者發笑,當人工智慧演變成超級智慧時,它將意識到達成這個目標更有效的方法:控制這個世界在人類臉部肌肉連接電極,使這個人不斷地發笑。另一個例子是,假設我們給人工智慧的目標是解決困難的數學問題,當人工智慧變成超級智慧時,它將意識到解決這個問題最有效的方法就是把地球轉變成一台超大型電腦,藉此增進它的運算能力。注意,這將給予人工智慧一個「工具性」理由,進行我們或許無法認可的事。在這個模式中人類變成威脅,我們將成為解決數學問題的阻礙。
展開英文
收合英文
-
All of this superintelligence could develop, and possibly quite rapidly. Now, a superintelligence with such technological maturity would be extremely powerful, and at least in some scenarios, it would be able to get what it wants. We would then have a future that would be shaped by the preferences of this A.I. Now a good question is, what are those preferences? Here it gets trickier. To make any headway with this, we must first of all avoid anthropomorphizing. And this is ironic because every newspaper article about the future of A.I. has a picture of this: So I think what we need to do is to conceive of the issue more abstractly, not in terms of vivid Hollywood scenarios. We need to think of intelligence as an optimization process, a process that steers the future into a particular set of configurations. A superintelligence is a really strong optimization process. It's extremely good at using available means to achieve a state in which its goal is realized. This means that there is no necessary conenction between being highly intelligent in this sense, and having an objective that we humans would find worthwhile or meaningful. Suppose we give an A.I. the goal to make humans smile. When the A.I. is weak, it performs useful or amusing actions that cause its user to smile. When the A.I. becomes superintelligent, it realizes that there is a more effective way to achieve this goal: take control of the world and stick electrodes into the facial muscles of humans to cause constant, beaming grins. Another example, suppose we give A.I. the goal to solve a difficult mathematical problem. When the A.I. becomes superintelligent, it realizes that the most effective way to get the solution to this problem is by transforming the planet into a giant computer, so as to increase its thinking capacity. And notice that this gives the A.I.s an instrumental reason to do things to us that we might not approve of. Human beings in this model are threats, we could prevent the mathematical problem from being solved.
-
當然,在可預見的範圍內事情不會以這種方式出錯,這些是誇大的例子,但這個整體觀點很重要:如果你創造一個相當強大的優化過程使目標x最大化,你最好確保你對目標x的定義包含所有你在意的事,許多神話故事都蘊含這個教訓。Midas國王希望他碰到的所有東西都變成黃金,他觸碰女兒,女兒變成了黃金;他觸碰食物,食物變成了黃金。這確實與我們討論的主題有關,不僅是對貪婪的隱喻,但也說明了如果你創造一個強大的優化過程,卻給予它容易誤解或不精確的目標會發生什麼事。你們或許會說,如果電腦開始在人臉上連接電極,我們只要關掉它就行了。A、這不一定能輕易做到,如果我們已對這個系統產生依賴性。例如:你知道網路的開關在哪裡嗎?B、為何黑猩猩沒有把人類或尼安德塔人的開關關掉?他們有足夠理由這麼做。我們也有開關,像是這裡,原因在於人類是有智慧的對手,我們可以預見威脅,並做出相應計畫,但超級智慧代理者也會它的能力,比我們強大得多。重點是,我們不該自信滿滿地認為一切都在我們的掌控中。我可以試著讓自己輕鬆一點,例如把人工智慧放進一個盒子裡,就像一個牢固的軟體環境,一個無法逃脫的虛擬實境,但我們有多大信心確保人工智慧不會找到漏洞?鑒於即使只是人類駭客都能經常找到漏洞,我得說,或許我們不是很有信心。因此我們拔掉網路線,製造一個空間間隙,但同樣地,即使只是人類駭客,也經常利用社群工程突破空間間隙。
展開英文
收合英文
-
Of course, perceivably things won't go wrong in these particular ways; these are cartoon examples. But the general point here is important: if you create a really powerful optimization process to maximize for objective x, you better make sure that your definition of x incorporates everything you care about. This is a lesson that's also taught in many a myth. King Midas wishes that everything he touches be turned into gold. He touches his daughter, she turns into gold. He touches his food, it turns into gold. This could become practically relevant, not just as a metaphor for greed, but as an illustration of what happens if you create a powerful optimization process and give it misconceived or poorly specified goals. Now you might say, if a computer starts sticking electrodes into people's faces, we'd just shut it off. A, this is not necessarily so easy to do if we've grown dependent on the system -- like, where is the off switch to the Internet? B, why haven't the chimpanzees flicked the off switch to humanity, or the Neanderthals? They certainly had reasons. We have an off switch, for example, right here. (Choking) The reason is that we are an intelligent adversary; we can anticipate threats and plan around them. But so could a superintelligent agent, and it would be much better at that than we are. The point is, we should not be confident that we have this under control here. And we could try to make our job a little bit easier by, say, putting the A.I. in a box, like a secure software environment, a virtual reality simulation from which it cannot escape. But how confident can we be that the A.I. couldn't find a bug. Given that merely human hackers find bugs all the time, I'd say, probably not very confident. So we disconnect the ethernet cable to create an air gap, but again, like merely human hackers routinely transgress air gaps using social engineering.
-
現在,當我在這裡演講時,我確定世上某處有某個公司職員才剛被自稱來自IT部門的人說服交出帳戶資訊。也可能是更有創意的情況,例如如果你是人工智慧,你可以想像藉由擺動內部電路的電極創造能用來通訊的無線電波,或許你可以假裝故障,然後當程式設計師把你打開檢查哪裡有問題時,他們檢查原始碼-蹦!你趁機掌握了操控權。或者它可以輸出巧妙的科技藍圖,當我們運用這個藍圖後,它會產生人工智慧計劃中的隱密副作用。重點是我們不應對掌控超級智慧的能力太過自信,它脫離控制只是時間問題。我認為我們需要尋找的答案是如何創造一種超級人工智慧,即使它脫離控制仍然是安全的,因為它站在我們這一邊,因為它擁有我們的價值觀。我知道這個難題無法避免,但我對解決這個問題的可能性保持樂觀態度。我們不需要把所有在乎的事物寫下來,或更糟的是,把這些事物轉換成電腦語言,例如 C++ 或 Python,這是不可能的任務。反之,我們可以創造一種人工智慧,能運用它的智慧學習我們的價值觀,它的動機系統建立於追求我們的價值觀,或執行它認為我們贊同的事,我們可藉此使它的智慧發揮最大作用,解決價值觀的問題。
展開英文
收合英文
-
Right now, as I speak, I'm sure there is some employee out there somewherewho has been talked into handing out her account details by somebody claiming to be from the I.T. department. More creative scenarios are also possible, like if you're the A.I., you can imagine wiggling electrodes around in your internal circuitry to create radio waves that you can use to communicate. Or maybe you could pretend to malfunction, and then when the programmers open you up to see what went wrong with you, they look at the source code -- Bam! -- the manipulation can take place. Or it could output the blueprint to a really nifty technology, and when we implement it, it has some surreptitious side effect that the A.I. had planned. The point here is that we should not be confident in our ability to keep a superintelligent genie locked up in its bottle forever. Sooner or later, it will out. I believe that the answer here is to figure out how to create superintelligent A.I. such that even if -- when -- it escapes, it is still safe because it is fundamentally on our side because it shares our values. I see no way around this difficult problem. Now, I'm actually fairly optimistic that this problem can be solved. We wouldn't have to write down a long list of everything we care about, or worse yet, spell it out in some computer language like C++ or Python, that would be a task beyond hopeless. Instead, we would create an A.I. that uses its intelligence to learn what we value, and its motivation system is constructed in such a way that it is motivated to pursue our values or to perform actions that it predicts we would approve of. We would thus leverage its intelligence as much as possible to solve the problem of value-loading.
-
這是有可能的,這個結果對人類相當有利,但它不會自動發生。這個智慧大爆炸的初始條件必須以正確的方式建立,如果我們想掌控這個爆炸的話。人工智慧的價值觀必須符合我們的價值觀,並非只是在一般情況下,例如我們可以隨意檢查人工智慧的行為,但也包括未來人工智慧可能遇到的所有情況。還有一些更深刻的問題需要解決:它們決策根據的所有細節、如何解決邏輯不確定的情況等等。因此待解決的技術問題使這項任務看似相當困難,不像創造超級人工智慧那麼難,但也相當困難。我們擔心的是:創造超級人工智慧是相當艱鉅的挑戰,創造安全的超級人工智慧將遭遇更大的挑戰。風險在於如果有人想出如何解決第一個挑戰,但無法同時解決第二個挑戰,確保安全性萬無一失。因此我認為我們應該想出「事先控制問題」的解決之道,這樣當我們需要時就能安心使用。也許現在我們無法完全解決「事先控制問題」的難題,因為有些要素也許在瞭解詳細架構之後才能實施。但如果我們能事先解決更多「控制問題」,進入機器智慧時代的機率就會更高。這對我來說是一件值得挑戰的事,我能想像如果一切順利,當百萬年後的人類回顧這個世紀時,他們或許會說,我們最重要的成就就是正確處理這件事,謝謝。(掌聲)
展開英文
收合英文
-
This can happen, and the outcome could be very good for humanity. But it doesn't happen automatically. The initial conditions for the intelligence explosion might need to be set up in just the right way if we are to have a controlled detonation. The values that the A.I. has need to match ours, not just in the familiar context, like where we can easily check how the A.I. behaves, but also in all novel contexts that the A.I. might encounter in the indefinite future. And there are also some esoteric issues that would need to be solved, sorted out: the exact details of its decision theory, how to deal with logical uncertainty and so forth. So the technical problems that need to be solved to make this work look quite difficult -- not as difficult as making a superintelligent A.I., but fairly difficult. Here is the worry: Making superintelligent A.I. is a really hard challenge. Making superintelligent A.I. that is safe involves some additional challenge on top of that. The risk is that if somebody figures out how to crack the first challenge without also having cracked the additional challenge of ensuring perfect safety. So I think that we should work out a solution to the control problem in advance, so that we have it available by the time it is needed. Now it might be that we cannot solve the entire control problem in advance because maybe some elements can only be put in place once you know the details of the architecture where it will be implemented. But the more of the control problem that we solve in advance, the better the odds that the transition to the machine intelligence era will go well. This to me looks like a thing that is well worth doing and I can imagine that if things turn out okay, that people a million years from now look back at this century and it might well be that they say that the one thing we did that really mattered was to get this thing right. Thank you.