Placeholder Image

Subtitles section Play video

  • the AI race between China and the U.S. and what's at stake?

    中國和美國之間的人工智能競賽及其利害關係?

  • Okay so first of all China has a lot of disadvantages in competing with the U.S.

    首先,中國在與美國競爭時有很多劣勢。

  • Number one is the fact that they don't get access to all the hardware that we have access to here.

    第一,他們無法使用我們這裡的所有硬件。

  • So they're kind of working with lower-end GPUs than us.

    是以,他們使用的 GPU 比我們低端。

  • It's almost like working with the previous generation GPUs crappily.

    這幾乎就像是在蹩腳地使用上一代 GPU。

  • So and the fact that the bigger models tend to be more smarter naturally puts them at a disadvantage.

    是以,更大的機型往往更聰明,這自然使它們處於劣勢。

  • But the flip side of this is that necessity is the mother of invention.

    但反過來說,必要是發明之母。

  • Because they had to go figure out workarounds, they actually ended up building something a lot more efficient.

    因為他們不得不去想變通的辦法,所以實際上他們最終建造的東西效率要高得多。

  • It's like saying, hey look you guys really got to get a top-notch model and I'm not going to give you resources.

    這就好比說,嘿,你們真的需要一個頂級模特,但我不會給你們資源。

  • I mean figure out something, right?

    我的意思是想辦法,對吧?

  • Unless it's impossible, unless it's mathematically possible to prove that it's impossible to do so, you can always try to like come up with something more But that is likely to make them come up with a more efficient solution than America.

    除非這是不可能的,除非在數學上可以證明這是不可能的,否則你總可以試著想出一些更有可能讓他們想出比美國更有效解決方案的辦法。

  • And of course they've open sourced it so we can still adopt something like that here.

    當然,他們已經將其開源,所以我們仍然可以在這裡採用類似的東西。

  • But that kind of talent they're building to do that will become an edge for them over time, right?

    但他們正在培養的這種人才,隨著時間的推移,會成為他們的優勢,對嗎?

  • The leading open source model in America is Meta's Lama family.

    美國領先的開源模式是 Meta 的喇嘛家族。

  • It's really good.

    真的很不錯。

  • It's kind of like a model that you can run on a computer.

    它有點像一個可以在電腦上運行的模型。

  • But even though it got pretty close to GPT-4 and Asana at the time of its release, the model that was closest in quality was a giant 405B, not the 70B that you could run on your computer.

    不過,儘管它在發佈時已經非常接近 GPT-4 和 Asana,但在品質上最接近的型號是巨大的 405B,而不是可以在電腦上運行的 70B。

  • And so that was still not a small, cheap, fast, efficient, open source model that rivaled the most powerful closed models from Okinawa and Tropic.

    是以,這仍然不是一個小型、廉價、快速、高效的開源模型,無法與沖繩和熱帶最強大的封閉模型相媲美。

  • Nothing from America.

    沒有來自美國的東西。

  • Nothing from Mistral either.

    Mistral 也沒有任何消息。

  • And then these guys come out with like a crazy model that's like 10x cheaper in API pricing than GPT-40 and 15x cheaper than Sonnet, I believe.

    然後,這些傢伙又推出了一個瘋狂的型號,其 API 價格比 GPT-40 便宜 10 倍,比 Sonnet 便宜 15 倍。

  • Really fast, 60 tokens per second.

    速度非常快,每秒 60 塊令牌。

  • And pretty much equal or better in some benchmarks than worse than some others, but like roughly in that ballpark of 4OS quality.

    在某些基準測試中,它的性能相當或更好,而在其他一些基準測試中,它的性能則較差,但大致與 4OS 的品質相當。

  • And they did it all with like approximately just 2048H800 GPUs, which is actually equivalent to like somewhere around 1500 or 1500H100 GPUs.

    他們使用大約 2048H800 GPU 完成了這一切,這實際上相當於大約 1500 或 1500H100 GPU。

  • That's like 20 to 30x lower than the amount of GPUs that GPT-4 is usually trained on.

    這比 GPT-4 通常使用的 GPU 數量要低 20 到 30 倍。

  • And they did roughly $5 million in total compute budget.

    他們的計算預算總額約為 500 萬美元。

  • They did it with so little money and such an amazing model.

    他們用如此少的錢和如此出色的模型做到了這一點。

  • Gave it away for free.

    免費贈送。

  • Wrote a technical paper.

    撰寫了一篇技術論文。

  • And definitely it makes us all question like, okay, like if we have the equivalent of DOGE for like model training, this is an example of that, right?

    這肯定會讓我們產生這樣的疑問:好吧,如果我們有等同於 DOGE 的模型訓練方法,這不就是一個例子嗎?

  • Right.

  • Efficiency is what you're getting at.

    效率就是你要表達的意思。

  • So fraction of the price, fraction of the time, dumbed down GPUs essentially.

    是以,只需幾分之一的價格,幾分之一的時間,基本上就是傻瓜式 GPU。

  • What was your surprise when you understood what they had done?

    當你明白他們的所作所為時,你有什麼驚訝?

  • So my surprise was that when I actually went through the technical paper, the amount of clever solutions they came up with.

    是以,讓我感到驚訝的是,當我真正翻閱技術文件時,他們提出了大量巧妙的解決方案。

  • First of all, they trained a experts model.

    首先,他們訓練了一個專家模型。

  • It's not that easy to train.

    訓練沒那麼容易

  • There's a lot of like, the main reason people find it difficult to catch up with OpenAI, especially the MOE architecture, is that there's a lot of irregular loss spikes.

    人們發現 OpenAI(尤其是 MOE 架構)難以追趕的主要原因是,有很多不規則的損失峰值。

  • The numerics are not stable.

    數值不穩定。

  • So often like you've got to restart the training checkpoint again.

    所以,你經常要重新開始訓練檢查點。

  • And a lot of infrastructure needs to be built for that.

    為此需要建設大量的基礎設施。

  • And they came up with very clever solutions to balance that without adding additional hacks.

    他們想出了非常巧妙的辦法,在不增加額外黑客的情況下實現了平衡。

  • And they also figured out 8-bit training, at least for some of the numerics.

    他們還想出了 8 位訓練方法,至少對某些數字進行了訓練。

  • And they cleverly figured out which has to be in higher precision, which has to be in lower precision.

    他們巧妙地算出了哪個精度要高,哪個精度要低。

  • And to my knowledge, I think floating point 8 training is not that well understood.

    據我所知,我認為浮點 8 訓練還沒有得到很好的理解。

  • Most of the training in America is still running.

    美國的大部分培訓仍在進行。

  • And maybe OpenAI and some people are trying to explore that, but it's pretty difficult to get it right.

    也許 OpenAI 和一些人正在嘗試探索這一點,但要做到這一點非常困難。

  • So because necessities, I'm going to mention, because they don't have that much memory, that many GPUs, they figured out a lot of numerical stability stuff that makes their work.

    是以,我想說的是,因為他們沒有那麼多內存和 GPU,所以他們想出了很多數值穩定性的方法,使他們的工作變得更加穩定。

  • And they claimed in the paper that the majority of the training was stable, which means they can always rerun those training runs again on more data or better data.

    他們在論文中聲稱,大部分訓練是穩定的,這意味著他們隨時可以在更多數據或更好的數據上重新運行這些訓練運行。

  • And then it only trained for 60 days.

    然後只培訓 60 天。

  • So it's pretty amazing.

    所以,這非常了不起。

  • It's nice to say you were surprised.

    很高興你能說你很驚訝。

  • So I was definitely surprised.

    所以,我絕對是大吃一驚。

  • Usually the wisdom, or I would say the myth, is that Chinese are just good at copying.

    通常,人們認為中國人只是善於模仿,或者說是神話。

  • So if we stop writing research papers in America, if we stop describing the details of our infrastructure, our architecture, and stop open sourcing, they're not going to be able to catch up.

    是以,如果我們停止在美國撰寫研究論文,如果我們停止描述我們的基礎設施和架構的細節,停止開放源代碼,他們將無法迎頭趕上。

  • But the reality is, some of the details in DeepSea 3 are so good that I wouldn't be surprised if Meta took a look at it and incorporated some of that in Llamas 4.

    但實際情況是,《深海 3》中的一些細節非常出色,如果 Meta 借鏡了《深海 3》,並在《傲骨賢妻 4》中加入一些細節,我也不會感到驚訝。

  • I wouldn't necessarily say copy.

    我不一定說是抄襲。

  • It's all sharing science, engineering.

    這都是科學和工程學的共享。

  • But the point is, it's changing.

    但問題是,情況正在發生變化。

  • It's not like China is copycat.

    中國又不是模仿者。

  • They're also innovating.

    他們還在不斷創新。

  • We don't know exactly the data that it was trained on, even though it's open source.

    儘管它是開源的,但我們並不清楚其訓練所依據的數據。

  • We know some of the ways and things it was trained on, but not everything.

    我們知道它的一些訓練方式和訓練內容,但不是全部。

  • And there's this idea that it was trained on public chat GPT outputs, which would mean it just was copied.

    還有一種說法是,它是在公開哈拉 GPT 輸出上訓練出來的,這意味著它只是被複制了。

  • But you're saying it goes beyond that.

    但你說的不止於此。

  • There's real innovation.

    有真正的創新。

  • Yeah, look, they've trained it on 14.8 trillion tokens.

    是啊,你看,他們已經在 14.8 萬億個代幣上進行了訓練。

  • The internet has so much chat GPT.

    互聯網上有很多哈拉 GPT。

  • If you actually go to any LinkedIn post or X post now, most of the comments are written by AI.

    如果你現在真的去看 LinkedIn 上的帖子或 X 帖子,大部分評論都是人工智能寫的。

  • You can just see it.

    你可以看到它。

  • People are just trying to write.

    人們只是在努力寫作。

  • Even with an X, there's a Grok tweet enhancer.

    即使是 X,也有一個 Grok 鳴叫增強器。

  • Or in LinkedIn, there's an AI enhancer.

    或者在 LinkedIn 中,有一個人工智能增強器。

  • Or in Google Docs and Word, there are AI tools to rewrite your stuff.

    或者在谷歌文檔和 Word 中,也有人工智能工具來改寫你的內容。

  • So if you do something there and copy-paste it somewhere on the internet, it's naturally going to have some elements of a chat GPT-like training.

    是以,如果你在那裡做了一些事情,然後複製粘貼到互聯網上的某個地方,自然會有一些類似哈拉 GPT 培訓的元素。

  • And there's a lot of people who don't even bother to strip away that I'm a language model part.

    有很多人甚至懶得剝去我是語言模型的部分。

  • So they just paste it somewhere.

    所以他們就把它粘貼在某個地方。

  • It's very difficult to control for this.

    這一點很難控制。

  • I think XAI has spoken about this too.

    我想 XAI 也談到過這個問題。

  • So I wouldn't disregard their technical accomplishment just because for some prompts, like, who are you?

    是以,我不會因為某些提示,比如你是誰?

  • Or which model are you in response to that?

    或者說,你是哪種型號?

  • It doesn't even matter in my opinion.

    在我看來,這根本不重要。

  • For a long time, we thought, I don't know if you agreed with us, China was behind in AI.

    在很長一段時間裡,我們都認為--我不知道你是否同意--中國在人工智能方面落後了。

  • What does this do to that race?

    這對比賽有什麼影響?

  • Can we say that China is catching up or hasn't caught up?

    我們能說中國正在追趕還是沒有追趕嗎?

  • I mean, if we say the meta is catching up to open AI and entropic, if you make that claim, then the same claim can be made for China catching up to America.

    我的意思是,如果我們說元語言正在追趕開放式人工智能和熵,如果你這麼說,那麼同樣的說法也可以適用於中國追趕美國。

  • A lot of papers from China that have tried to replicate O1.

    中國有很多論文試圖複製 O1。

  • In fact, I saw more papers from China after O1 announcement that tried to replicate it than from America.

    事實上,在 O1 公佈後,我看到中國試圖複製 O1 的論文比美國還多。

  • And the amount of compute DeepSeq has access to is roughly similar to what PhD students in the US have access to.

    DeepSeq 所能使用的計算量與美國博士生所能使用的計算量大致相當。

  • This is not meant to criticize others, even for ourselves.

    這並不是要責備別人,即使是責備我們自己。

  • For Perplexity, we decided not to train models because we thought it's a very expensive thing.

    對於 Perplexity,我們決定不訓練模型,因為我們認為這是一件非常昂貴的事情。

  • And we thought there's no way to catch up with the rest.

    而我們卻認為沒有辦法趕上其他人。

  • Will you incorporate DeepSeq into Perplexity?

    你們會將 DeepSeq 納入 Perplexity 嗎?

  • We already are beginning to use it.

    我們已經開始使用它。

  • I assume they have an API and they have open source of AI so we can host it ourselves too.

    我想他們應該有一個應用程序接口,而且他們還開放了人工智能的源代碼,這樣我們也可以自己託管。

  • And it's good to try to start using that because it actually allows us to do a lot of the things at lower cost.

    嘗試開始使用這種方法是件好事,因為它實際上可以讓我們以更低的成本完成很多事情。

  • But what I'm kind of thinking is beyond that, it's just like, okay, these guys actually could train such a great model with a good team.

    但我的想法是,除了這一點之外,這些人其實還可以和一支優秀的團隊一起訓練出如此出色的模型。

  • And there's no excuse anymore for companies in the US, including ourselves, to not try to do something like that.

    包括我們自己在內的美國公司再也沒有藉口不嘗試這樣做了。

  • You hear a lot in public from a lot of thought leaders and generative AI, on the research side, on the entrepreneurial side.

    在公開場合,你會聽到很多思想領袖和生成式人工智能的觀點,無論是在研究方面,還是在創業方面。

  • Elon Musk and others say that China can't catch up.

    埃隆-馬斯克等人說,中國趕不上。

  • The stakes are too big, the geopolitical stakes.

    地緣政治的賭注太大了。

  • Whoever dominates AI is going to dominate the economy, dominate the world.

    誰主宰了人工智能,誰就能主宰經濟,主宰世界。

  • It's been talked about in those massive terms.

    我們一直在談論這些大型活動。

  • Are you worried about what China proved it was able to do?

    你擔心中國證明了自己的能力嗎?

  • Firstly, I don't know if Elon ever said China can't catch up.

    首先,我不知道埃隆是否說過中國趕不上。

  • I'm not aware of that.

    我不知道。

  • Just the threat of China.

    只是中國的威脅。

  • He's only identified the threat of letting China.

    他只指出了中國的威脅。

  • Sam Altman has said similar things.

    薩姆-奧特曼也說過類似的話。

  • We can't let China win.

    我們不能讓中國贏。

  • I think you've got to decouple what someone like Sam says to what is in his self interest.

    我認為,你必須把薩姆這樣的人所說的話與他自身的利益脫鉤。

  • My point is, whatever you did to not let them catch up didn't even matter.

    我想說的是,不管你做了什麼,不讓他們追上來都沒有關係。

  • They ended up catching up anyway.

    結果他們還是趕上了。

  • Necessity is the mother of invention, like you said.

    正如你所說,需要是發明之母。

  • Exactly.

    沒錯。

  • What's more dangerous than trying to do all the things to not let them catch up?

    還有什麼比竭盡全力不讓他們趕上更危險的呢?

  • What's more dangerous is they have the best open source model, and all the American developers are building on that.

    更危險的是,他們擁有最好的開源模式,所有的美國開發者都在此基礎上進行開發。

  • That's more dangerous, because then they get to own the mindshare, the ecosystem, the entire American AI ecosystem.

    這就更危險了,因為這樣一來,他們就能掌握整個美國人工智能生態系統的思想份額和生態系統。

  • In general, it's known that once open source is caught up or improved over closed source software, all developers migrate to that.

    一般來說,眾所周知,一旦開放源碼軟件趕上或改進了封閉源碼軟件,所有開發人員都會遷移到開放源碼軟件。

  • It's historically known.

    這在歷史上是眾所周知的。

  • When LLAMA was being built and becoming more widely used, there was this question, should we trust Zuckerberg?

    當 LLAMA 建立起來並被越來越廣泛地使用時,人們不禁要問:我們應該相信扎克伯格嗎?

  • But now the question is, should we trust China?

    但現在的問題是,我們應該相信中國嗎?

  • Should we trust open source?

    我們應該信任開放源代碼嗎?

  • It's not about who is in Zuckerberg.

    這與扎克伯格的身份無關。

  • Does it matter then if it's Chinese, if it's open source?

    那麼,它是否是中國的,是否是開源的,這重要嗎?

  • It doesn't matter in the sense that you still have full control.

    這並不重要,因為你仍然可以完全控制。

  • You run it as your own weights on your own computer.

    您可以在自己的電腦上將其作為自己的權重運行。

  • You are in charge of the model.

    你負責模型。

  • It's not a great look for our own talent to rely on software built by others.

    依賴他人開發的軟件對我們自己的人才來說並不光彩。

  • There's always a point where open source can stop being open source too.

    開放源代碼也有停止開放源代碼的時候。

  • The licenses are very favorable today, but over time, they can always change the license.

    今天的許可證非常有利,但隨著時間的推移,他們可以隨時更改許可證。

  • It's important that we actually have people here in America building, and that's why Meta is so important.

    重要的是,我們要讓美國人民真正參與到建設中來,這也是 Meta 如此重要的原因。

  • I still think Meta will build a better model than DeepSecret 3 and open source, and what they call LLAMA 4 or 3.something.

    我仍然認為,Meta 將建立一個比 DeepSecret 3 和開源,以及他們所謂的 LLAMA 4 或 3.something 更好的模型。

  • It doesn't matter.

    沒關係。

  • What is more key is that we don't try to focus all our energy on banning them, stopping them, and just try to out-compete and win.

    更關鍵的是,我們不要試圖把所有精力都放在禁止它們、阻止它們上,而要努力超越它們,贏得勝利。

  • That's just the American way of doing things.

    這就是美國人的做事方式。

  • Just be better.

    做得更好

  • It feels like we hear a lot more about these Chinese companies who are developing in a similar way, a lot more efficiently, a lot more cost-effectively, right?

    我們似乎聽到了更多關於這些中國公司的報道,它們以類似的方式發展,效率更高,成本效益更好,對嗎?

  • Again, it's hard to fake scarcity, right?

    還是那句話,稀缺是很難偽造的,不是嗎?

  • If you raise $10 billion and you are decided to spend 80% of it on a computer cluster, it's hard for you to come up with the exact same solution that $5 million would do, and there's no need to berate those who are putting more money.

    如果你籌集了 100 億美元,卻決定把其中的 80% 用在計算機集群上,那麼你很難想出與 500 萬美元完全相同的解決方案,也就沒有必要斥責那些投入更多資金的人。

  • They're trying to do it as fast as they can.

    他們想盡快完成任務。

  • When we say open source, there are so many different versions.

    當我們說開放源代碼時,有許多不同的版本。

  • Some people criticize Meta for not publishing everything, and even DeepSpeak itself isn't totally transparent.

    有人責備 Meta 沒有公佈所有資訊,甚至 DeepSpeak 本身也不是完全透明的。

  • Sure.

    當然。

  • You can go to the limits of open source and say, I should exactly be able to replicate your training run, but first of all, how many people even have the resources to do that?

    你可以挑戰開源的極限,說我應該可以完全複製你的訓練運行,但首先,有多少人有這樣的資源?

  • I think the amount of detail they've shared in the technical report, actually Meta did that too, by the way.

    我認為他們在技術報告中分享了大量細節,事實上,Meta 也是這麼做的。

  • Meta's Lama 3.3 technical report is incredibly detailed and very great for science.

    Meta 的喇嘛 3.3 技術報告非常詳細,對科學非常有幫助。

  • The amount of detailed data these people are sharing is already a lot more than what the other companies are doing right now.

    這些人共享的詳細數據量已經遠遠超過了其他公司現在所做的。

  • When you think about how much it costs DeepSpeak to do this, less than $6 million, think about what OpenAI has spent to develop GPT models.

    想想 DeepSpeak 做這件事的成本(不到 600 萬美元),再想想 OpenAI 開發 GPT 模型的成本。

  • What does that mean for the closed source model, ecosystem trajectory, momentum?

    這對閉源模式、生態系統軌跡和發展勢頭意味著什麼?

  • What does it mean for OpenAI?

    這對 OpenAI 意味著什麼?

  • It's very clear that we'll have an open source version of 4.0, or even better than that, and much cheaper than that, open source, completely in this year.

    很明顯,我們將在今年完全推出 4.0 的開源版本,甚至比它更好,比它更便宜的開源版本。

  • Made by OpenAI?

    由 OpenAI 製造?

  • Probably not.

    可能不會。

  • Most likely not.

    很可能不會。

  • I don't think they care if it's not made by them.

    我認為,如果不是他們生產的,他們也不在乎。

  • I think they've already moved to a new paradigm called the O1 family of models.

    我認為他們已經轉向了一種新的模式,即 O1 系列機型。

  • Ilya Sutskiy came and said, pre-training is a wall.

    伊利亞-蘇茨基來了,他說,訓練前是一堵牆。

  • He didn't exactly use the word, but he clearly said the age of pre-training is a wall.

    他沒有確切地使用這個詞,但他明確地說,訓練前的年齡是一堵牆。

  • Many people have said that.

    很多人都這麼說過。

  • That doesn't mean scaling is a wall.

    這並不意味著縮放就是一堵牆。

  • I think we're scaling on different dimensions now.

    我想,我們現在是在不同的維度上進行縮放。

  • The amount of time the model spends thinking at this time, reinforcement learning, trying to make the model, if it doesn't know what to do for a new prompt, it'll go and reason and collect data and interact with the world, use a bunch of tools.

    如果模型對新的提示不知道該怎麼做,它就會去推理、收集數據並與世界互動,使用一堆工具。

  • I think that's where things are headed.

    我認為這就是事情的發展方向。

  • I feel like OpenAI is more focused on that right now.

    我覺得 OpenAI 現在更專注於此。

  • Instead of just a bigger, better model of reasoning capacity.

    而不僅僅是一個更大、更好的推理能力模型。

  • But didn't you say that DeepSeek is likely to turn their attention to reasoning?

    但你不是說 DeepSeek 可能會把注意力轉向推理嗎?

  • A hundred percent.

    百分之百。

  • I think they will.

    我想他們會的。

  • That's why I'm pretty excited about what they'll produce next.

    這就是為什麼我對他們的下一部作品相當期待。

  • I guess my question is, what's OpenAI's moat now?

    我想我的問題是,OpenAI 現在的護城河是什麼?

  • I still think that no one else has produced a system similar to the O1 yet, exactly.

    我仍然認為,確切地說,還沒有人生產出與 O1 類似的系統。

  • I know that there's debates about whether O1 is actually worth it.

    我知道,關於 O1 是否真的值得的爭論一直存在。

  • Maybe a few prompts, it's really better.

    也許再加上一些提示,效果會更好。

  • But most of the time, it's not producing any differentiated output from Sonnet.

    但在大多數情況下,它與 Sonnet 沒有任何區別。

  • But at least the results they showed in O3 where they had competitive coding performance, almost like an AI software engineer level.

    但至少,他們在 O3 中表現出的編碼性能很有競爭力,幾乎達到了人工智能軟件工程師的水準。

  • Isn't it just a matter of time before the internet is filled with reasoning data?

    互聯網上充斥著推理數據難道不是遲早的事嗎?

  • Again, it's possible.

    同樣,這也是可能的。

  • Nobody knows yet.

    還沒人知道。

  • So until it's done, it's still uncertain.

    是以,在完成之前,一切都還不確定。

  • So maybe that uncertainty is their moat, that no one else has the same reasoning capability yet.

    所以,也許這種不確定性就是他們的護城河,因為還沒有人具備同樣的推理能力。

  • By the end of this year, will there be multiple players even in the reasoning arena?

    到今年年底,即使在推理領域也會有多個參與者嗎?

  • I absolutely think so.

    我絕對這麼認為。

  • So are we seeing the commoditization of large language models?

    那麼,我們是否看到了大型語言模型的商品化?

  • I think we'll see a similar trajectory, just like how in pre-training and post-training, that's our system for getting commoditized.

    我認為我們會看到類似的軌跡,就像訓練前和訓練後一樣,這就是我們的商品化系統。

  • Where this year will be a lot more commoditization there.

    今年的商品化程度會更高。

  • I think the reasoning kind of models will go through a similar trajectory, where in the beginning, one or two players, they know how to do it.

    我認為,推理類模型也會經歷類似的軌跡,一開始,一兩個玩家就知道怎麼做。

  • But over time, like...

    但隨著時間的推移,就像...

  • And who knows, right?

    誰知道呢,對吧?

  • Because opening AI could make another advancement to focus on.

    因為開放人工智能可以讓我們關注另一個進步。

  • But right now, reasoning is their moat.

    但現在,推理是他們的護城河。

  • But if advancements keep happening again and again and again, I think the meaning of the word advancement also loses some of its value, right?

    但是,如果進步一而再、再而三地發生,我想進步這個詞的意義也就失去了一些價值,對嗎?

  • Totally.

    完全是

  • Even now, it's very difficult, right?

    即使是現在,也很困難,對嗎?

  • Because there's pre-training advancements, and then we've moved into a different thing.

    因為有了訓練前的進步,我們就進入了另一個階段。

  • Yeah.

    是啊

  • So what is guaranteed to happen is whatever models exist today, that level of reasoning, that level of multimodal capability, in like 5 to 10x cheaper models, open source, all that's going to happen.

    是以,可以肯定的是,無論今天存在什麼樣的模型,這種推理水準、這種多模態能力水準,都會以比現在便宜 5 到 10 倍的開源模型出現。

  • It's just a matter of time.

    這只是時間問題。

  • What is unclear is if something like a model that reasons at test time will be extremely cheap enough that we can just all run it on our phones.

    目前還不清楚的是,類似測試時的模型是否會非常便宜,以至於我們都可以在手機上運行它。

  • I think that's not clear to me yet.

    我想我還不太清楚。

  • It feels like so much of the landscape has changed with what DeepSeq was able to prove.

    DeepSeq 所能證明的一切,讓人感覺很多事情都發生了變化。

  • Could you call it China's chat to DT moment?

    你能說這是中國與 DT 對話的時刻嗎?

  • Possible.

    可能。

  • I think it certainly probably gave them a lot of confidence that we're not really behind.

    我想這可能給了他們很大的信心,我們並沒有真的落後。

  • No matter what you do to restrict or compute, we can always figure out some workarounds.

    無論您如何限制或計算,我們總能想出一些變通辦法。

  • And yeah, I'm sure the team feels pumped about the results.

    是的,我相信隊員們一定對結果感到振奮。

  • How does this change the investment landscape?

    這將如何改變投資格局?

  • The hyperscalers that are spending tens of billions of dollars a year on CapEx, they just ramped it up huge.

    超大規模企業每年的資本支出高達數百億美元,他們剛剛大幅提高了資本支出。

  • And OpenAI and Anthrobic that are raising billions of dollars for GPUs, essentially, that what DeepSeq told us is you don't need.

    而正在為 GPU 籌集數十億美元資金的 OpenAI 和 Anthrobic,從本質上講,DeepSeq 告訴我們的是你並不需要。

  • You don't necessarily need that.

    你不一定需要這樣。

  • Yeah.

    是啊

  • I mean, look, I think it's very clear that they're going to go even harder on reasoning because they understand that whatever they were building in the previous two years is getting extremely cheap.

    我的意思是,你看,我認為很明顯,他們會在推理上更加努力,因為他們明白,無論他們在前兩年建造了什麼,都變得非常便宜。

  • But it doesn't make sense to go justify raising that amount.

    但是,去證明提高這個數額是合理的並沒有意義。

  • Is the spending proposition the same?

    消費主張是否相同?

  • Do they need the same amount of high-end GPUs?

    它們需要相同數量的高端 GPU 嗎?

  • Or can you reason using the lower-end ones that DeepSeq is using?

    還是說,您可以使用 DeepSeq 使用的低端設備?

  • Again, it's hard to say no until proven it's not.

    還是那句話,在證明不是之前,很難說不是。

  • But I guess in the spirit of moving fast, you would want to use the high-end chips.

    但我想,本著快速發展的精神,你會希望使用高端芯片。

  • And you would want to move faster than your competitors.

    您希望比競爭對手走得更快。

  • I think the best talent still wants to work in the team that made it happen first.

    我認為,最優秀的人才仍然希望在最先實現夢想的團隊中工作。

  • There's always some glory to who did this actually, who's a real pioneer versus who's a fast follower, right?

    誰做了這件事,誰是真正的先驅,誰是快速的追隨者,總有一些榮耀,對嗎?

  • That was kind of like Sam Altman's tweet.

    這有點像山姆-奧特曼的推特。

  • Kind of veiled response to what DeepSeq has been able to.

    這是對 DeepSeq 所取得成果的一種含蓄迴應。

  • He kind of implied that they just copied and anyone can copy, right?

    他有點暗示說,他們只是抄襲,任何人都可以抄襲,對嗎?

  • Yeah.

    是啊

  • But then you can always say that everybody copies everybody in this field.

    不過,你也可以說,在這個領域,每個人都在模仿別人。

  • You can say Google did the transformer first.

    可以說,谷歌先做了變壓器。

  • It's not OpenAI.

    這不是 OpenAI。

  • And OpenAI just copied it.

    而 OpenAI 剛剛複製了它。

  • Google built the first large language models.

    谷歌建立了第一個大型語言模型。

  • They didn't prioritize it.

    他們沒有優先考慮這個問題。

  • But OpenAI did it in a prioritized way.

    但是,OpenAI 是按照優先順序來做的。

  • So you can say all this in many ways.

    所以,你可以用很多方式來表達這一切。

  • It doesn't matter.

    沒關係。

  • I remember asking you being like, why don't you want to build the model?

    我記得我問過你,你為什麼不想製作模型?

  • That's the glory.

    這就是榮耀。

  • And a year later, just one year later, you look very, very smart to not engage in that extremely expensive race that has become so competitive.

    而一年之後,僅僅一年之後,你就顯得非常非常聰明,沒有參與這場競爭已經變得如此激烈的極其昂貴的競賽。

  • And you kind of have this lead now in what everyone wants to see now, which is like real world applications, killer applications of generative AI.

    而你現在在大家都想看到的領域中處於領先地位,這就好比現實世界中的應用,生成式人工智能的殺手級應用。

  • Talk a little bit about that decision and how that's sort of guided you where you see perplexity going from here.

    請談一下這個決定,以及它是如何引導你將困惑帶向未來的。

  • Look, one year ago, I don't even think we had something like...

    你看,一年前,我甚至不認為我們有像...

  • This is what like 2024 beginning, right?

    這就像是 2024 年的開始,對嗎?

  • I feel like we didn't even have something like Sonic 3.5. We had GPT-4, I believe.

    我覺得我們甚至沒有類似 Sonic 3.5 的東西。 我記得我們有 GPT-4。

  • And it was kind of nobody else was able to catch up to it.

    其他人都沒能趕上它。

  • Yeah.

    是啊

  • But there was no multimodal, nothing.

    但沒有多模式,什麼都沒有。

  • And my sense was like, OK, if people with way more resources and way more talent cannot catch up, it's very difficult to play that game.

    我的感覺是,好吧,如果擁有更多資源和天賦的人都趕不上,那就很難玩這個遊戲了。

  • So let's play a different game.

    那我們就換一種玩法。

  • Anyway, people want to use these models.

    無論如何,人們都想使用這些模型。

  • And there's one use case of asking questions and getting accurate answers with sources, with real time information, accurate information.

    有一個用例是,提出問題並通過資訊來源、實時資訊和準確資訊獲得準確答案。

  • There's still a lot of work there to do outside the model and making sure the product works reliably, keep scaling it up to usage, keep building custom UIs.

    在模型之外還有很多工作要做,要確保產品可靠運行,不斷擴大使用規模,不斷構建自定義用戶界面。

  • There's a lot of work to do.

    還有很多工作要做。

  • And we focus on that.

    我們的重點就是這個。

  • And we would benefit from all the tailwinds of models getting better and better.

    我們將受益於模型越來越好的所有尾風。

  • That's essentially what happened.

    事情基本上就是這樣。

  • In fact, I would say Sonic 3.5 made our products so good in the sense that if you use Sonic 3.5 as the model choice within Perplexity, it's very difficult to find a hallucination.

    事實上,我認為 Sonic 3.5 使我們的產品如此出色,因為如果在 Perplexity 中使用 Sonic 3.5 作為模型選擇,就很難找到幻覺。

  • I'm not saying it's impossible, but it dramatically reduced the rate of hallucinations, which meant the problem of question answering, asking a question, getting an answer, doing a fact check, research, going and asking anything out there, because almost all the information is on the web, was such a big unlock.

    我並不是說這不可能,但它大大降低了幻覺的發生率,這意味著回答問題、提出問題、得到答案、進行事實核查、研究、去詢問外面的任何事情,都是一個巨大的難題,因為幾乎所有的資訊都在網絡上。

  • And that helped us grow 10x over the course of the year in terms of usage.

    在這一年中,我們的使用量增長了 10 倍。

  • And you've made huge strides in terms of users.

    你們在用戶方面取得了巨大進步。

  • And we hear on CNBC a lot.

    我們在 CNBC 上也經常聽到。

  • Big investors who are huge fans.

    大投資者,他們都是忠實粉絲。

  • Jensen Huang himself, right?

    黃健翔自己,對吧?

  • You mentioned that in his keynote the other night.

    你在他那天晚上的主題演講中提到了這一點。

  • He's a pretty regular user.

    他是一個非常固定的用戶。

  • He's not just saying it.

    他不只是說說而已。

  • He's actually a pretty regular user.

    實際上,他是一個非常固定的用戶。

  • So a year ago, we weren't even talking about monetization because you guys were just so new and you wanted to get yourselves out there and build some scale.

    一年前,我們甚至還沒有討論過貨幣化的問題,因為你們都是新手,你們想讓自己走出去,建立一定的規模。

  • But now you are looking at things like that, increasingly an ad model, right?

    但現在你們正在考慮這樣的事情,越來越多地採用廣告模式,對嗎?

  • Yeah, we're experimenting with it.

    是的,我們正在試驗。

  • I know there's some controversy on why should we do ads, whether you can have a truthful answer engine despite having ads.

    我知道,對於為什麼要做廣告,是否可以在有廣告的情況下仍能擁有一個真實的答案引擎,還存在一些爭議。

  • And in my opinion, we've been pretty proactively thoughtful about it, where we said, OK, as long as the answer is always accurate, unbiased, and not corrupted by someone's advertising budget, only you get to see some sponsored questions.

    在我看來,我們在這方面已經考慮得相當周到,我們說,好吧,只要答案始終準確、公正,不被某人的廣告預算所侵蝕,只有你能看到一些贊助商的問題。

  • And even the answers to the sponsored questions are not influenced by them.

    甚至連贊助問題的答案也不受它們的影響。

  • And questions are also not picked in a way where it's manipulative.

    問題的選取也沒有操縱性。

  • Sure, there are some things that the advertiser also wants, which is they want you to know about their brand.

    當然,廣告商也希望你瞭解他們的品牌。

  • And they want you to know the best parts of their brand.

    他們希望你瞭解其品牌的精華部分。

  • Just like how if you're introducing yourself to someone, you want them to see the best parts of you, right?

    就像你在向別人介紹自己時,希望對方看到你最好的一面,對嗎?

  • So that's all there.

    就這些了。

  • But you still don't know how to click on a sponsored question.

    但你還是不知道如何點擊贊助商問題。

  • You can ignore it.

    你可以置之不理。

  • And we are only charging them CPM right now.

    而我們現在只向他們收取 CPM。

  • So we ourselves are not even incentivized to make you click yet.

    所以,我們自己還沒有讓你點擊的動力。

  • So I think considering all this, we're actually trying to get it right long term instead of going the Google way of forcing you to click on links.

    是以我認為,考慮到這一切,我們實際上是在努力做好長期工作,而不是像谷歌那樣強迫你點擊鏈接。

  • I remember when people were talking about the commoditization of models a year ago, and you thought, oh, it was controversial.

    我記得一年前人們還在談論模型商品化的時候,你會想,哦,這是有爭議的。

  • But now it's not controversial.

    但現在這已經沒有爭議了。

  • It's kind of like that's happening.

    這有點像正在發生的事情。

  • You're keeping your eye on that.

    你要注意這一點。

  • It's smart.

    它很聰明。

  • By the way, we benefit a lot from model commoditization.

    順便說一句,我們從模型商品化中獲益良多。

  • Except we also need to figure out something to offer to the paid users, like a more sophisticated research agent that can do multi-step reasoning, go and do 15 minutes worth of searching, and give you an analysis, an analyst type of answer.

    不過,我們還需要為付費用戶提供一些東西,比如一個更復雜的研究代理,它可以進行多步推理,進行 15 分鐘的搜索,然後給你一個分析、分析師類型的答案。

  • All that's going to come.

    一切都會到來。

  • All that's going to stay in the product.

    所有這些都將留在產品中。

  • Nothing changes there.

    一切照舊。

  • But there's a ton of questions every free user asks day-to-day basis that needs to be quick, fast answers.

    但是,每個免費用戶每天都會提出大量問題,這些問題都需要得到快速的解答。

  • It shouldn't be slow.

    速度應該不慢。

  • And all that will be free, whether you like it or not.

    無論你喜歡與否,這一切都是免費的。

  • It has to be free.

    它必須是免費的。

  • That's what people are used to.

    這就是人們的習慣。

  • And that means figuring out a way to make that free traffic also monetizable.

    這意味著要想辦法讓免費流量也能賺錢。

  • So you're not trying to change user habits.

    所以,你並不是要改變用戶的習慣。

  • But it's interesting, because you are kind of trying to teach new habits to advertisers.

    但這很有趣,因為你是在向廣告商傳授新的習慣。

  • They can't have everything that they have in a Google Time Blue Links search.

    他們不可能擁有谷歌時間藍色鏈接搜索中的一切。

  • What's the response been from them so far?

    目前他們的反應如何?

  • Are they willing to accept some of the trade-offs?

    他們是否願意接受某些權衡?

  • Yeah.

    是啊

  • That's why they're trying stuff.

    這就是他們嘗試的原因。

  • Intuit is working with us.

    Intuit 正在與我們合作。

  • And then there's many other brands.

    還有許多其他品牌。

  • All these people are working with us to test.

    所有這些人都在與我們合作進行測試。

  • They're also excited about it.

    他們也為此感到興奮。

  • Everyone knows that, whether they like it or not, 5 to 10 years from now, most people are going to be asking AIs, most of the things, and not on the traditional search engine.

    每個人都知道,不管他們願不願意,從現在起的 5 到 10 年內,大多數人都會詢問人工智能,而不是在傳統的搜索引擎上。

  • Everybody understands that.

    每個人都明白這一點。

  • Everybody wants to be early adopters of the new platforms, new UX, and learn from it, and build things together.

    每個人都想成為新平臺、新用戶體驗的早期採用者,並從中學習,共同創造。

  • They're not viewing it as like, OK, you guys go figure out everything else.

    他們不會把它看成是 "好吧,你們去想別的辦法吧"。

  • And then we'll come later.

    然後我們再過來

  • I'm smiling, because it goes back perfectly to the point you made when you first sat down today, which is necessity is the mother of all invention, right?

    我笑了,因為這完全回到了你今天第一次坐下來時所說的觀點,即必要性是一切發明之母,對嗎?

  • And that's what advertisers are essentially looking at.

    而這正是廣告商所關注的。

  • The field is changing.

    該領域正在發生變化。

  • We have to learn to adapt with it.

    我們必須學會適應它。

  • OK, Arvind, I took up so much of your time.

    好吧,阿文德,我佔用了你這麼多時間。

  • Thank you so much for taking the time.

    非常感謝你抽出時間。

the AI race between China and the U.S. and what's at stake?

中國和美國之間的人工智能競賽及其利害關係?

Subtitles and vocabulary

Click the word to look it up Click the word to find further inforamtion about it

A2 訓練 中國 人工 智能 ai 使用

DeepSeek是在 "抄襲 "美國嗎? 它的出現對中美AI競爭的影響 DeepSeek | LLM |Open AI | 中國 |美國 |人工智能競爭 |開源模型 20250126金融汪 (DeepSeek是在“抄袭”美国吗?它的出现对中美AI竞争的影响 DeepSeek | LLM |Open AI | 中国 |美国 |人工智能竞争 |开源模型 20250126金融汪)

  • 6 1
    佛斯特 posted on 2025/01/29
Video vocabulary