开始:华尔街见闻
黄仁勋默示,莫得物理定律罢了AI数据中心扩展到百万芯片,我们当今不错将AI软件扩展到多个数据中心运行。我们仍是为能够在一个前所未有的水平上扩展计较作念好了准备,而且我们当今才刚刚运行。在畴昔十年,计较性能每年将翻倍或翻三倍,而动力需求每年将减少2-3倍,我称之为超摩尔定律弧线。
本周,英伟达CEO黄仁勋收受了《No Priors》节目主合手东说念主的采访,就英伟达的十年赌注、x.AI超等集群的快速发展、NVLink时刻改进等AI关系话题进行了一场深度对话。
黄仁勋默示,莫得任何物理定律不错封闭将AI数据中心扩展到一百万个芯片,尽管这是一个难题,多家大公司包括OpenAI、Anthropic、谷歌、Meta和微软等,都在争夺AI范围的带领地位,竞相攀高时刻的岑岭,但再行创造智能的潜在通告是如斯之大,以至于不可不去尝试。
摩尔定律曾是半导体行业发展的中枢法规,预测芯片的晶体管数量每两年会翻倍,从而带来性能的合手续升迁。关联词,跟着物理极限的接近,摩尔定律的速率运行放缓,芯片性能升迁的瓶颈逐渐清晰。
为了处置这一问题,英伟达将不同类型的处理器(如GPU、TPU等)皆集起来,通过并行处理来冲突传统摩尔定律的罢了。黄仁勋默示,畴昔10年,计较性能每年将翻一番或三倍,而动力需求每年将减少2-3倍,我称之为“超摩尔定律弧线”。
黄仁勋还提到,我们当今不错将AI软件扩展到多个数据中心:“我们仍是作念好准备,能够将计较扩展到前所未有的水平,而我们正处于这一范围的起步阶段。”
以下是黄仁勋讲话的亮点:
1.我们在畴昔10年进行了要紧的投资。我们正在投资基础设施,打造下一代AI计较平台。我们在软件、架构、GPU以及所有罢了AI开发所需的组件上都进行了投资。
2.摩尔定律,即晶体管数量每两年翻倍的预言,也曾是半导体行业的增长指南。关联词,跟着物理极限的接近,摩尔定律已不再能够单独推动芯片性能的升迁。为了处置这一问题,英伟达接纳了雷同于“异构计较”的方式,行将不同类型的处理器(如GPU、TPU等)皆集起来,通过并行处理来冲突传统摩尔定律的罢了。英伟达的时刻改进,如CUDA架构和深度学习优化,使得AI应用得以在超越摩尔定律的环境中高速运行。
3.我们推出了NVLink作为互连时刻,它使得多个GPU能够协同责任,每个GPU处理责任负载的不同部分。通过NVLink,GPU之间的带宽和通讯才智大幅升迁,使得数据中心能够扩展并扶持AI责任负载。
4.畴昔的AI应用需要动态和弹性强的基础设施,能够适合各式范围和类型的AI任务。因此,英伟达致力于于构建不错天真成就和高效运营的基础设施,讲理从中微型AI技俩到超大范围超等计较集群的需求。
5.构建AI数据中心的要害是要同期优化性能和遵循。在AI责任负载中,你需要巨大的电力,而散热成为一个巨大的问题。是以我们花了多量时间优化数据中心的瞎想和运营,包括冷却系统和电力遵循。
6.在硬件快速发展的配景下,保合手软件与硬件架构的兼容性显得尤为迫切。黄仁勋提到,我们必须确保我们的软件平台,如CUDA,不错跨代硬件使用。开发者不应当每次我们推出新芯片刻都被动重写代码。因此,我们确保保合手向后兼容,并让软件能够在我们开发的任何新硬件上高效运行。
7.我们正在确立一个超等集群,叫作念X.AI,它将成为宇宙上最大的AI超等计较平台之一。这个超等集群将提供扶持一些最操办勃勃的AI技俩所需的计较才智。这是我们推动AI前进的一大步。
8.扩展AI数据中心的一个大挑战是料理它们滥用的巨大动力。问题不单是是构建更大、更快的系统。我们还必须处理运行这些超大范围系统时面对的热量和电力挑战。为了应答这一切,需要改进的工程时刻来确保基础设施能够应答。
9.AI在芯片瞎想中的作用日益迫切,黄仁勋指出,AI仍是在芯片瞎想中施展着迫切作用。我们使用机器学习来匡助瞎想更高效的芯片,速率更快。这是我们瞎想下一代英伟达芯片的一个要害部分,并匡助我们构建专为AI责任负载优化的芯片。
10.英伟达市值的激增是因为我们能够将公司转型为AI公司。我们从一运行是GPU公司,但我们仍是转型成了AI计较公司,这一瞥型是我们市值增长的要害部分。AI时刻的需求正在飞速增长,我们处在一个能够讲理这一需求的成心位置。
11.具象化AI是指将AI与物理宇宙进行皆集。通过这种方式,AI不仅不错在编造环境中进行任务处理,还能在施行宇宙中进行决策并蔓延任务。具象化AI将推动智能硬件、自动驾驶等时刻的快速发展。
12.AI不单是是器用,它也不错成为‘编造职工’,匡助升迁责任遵循。AI能够在数据处理、编程、决策等范围替代或辅助东说念主类责任,进而窜改所有干事商场和责任方式。
13.AI将在科学与工程范围产生巨大影响,极度是在药物研发、场所研究、物理实验等范围。AI将匡助科学家处理多量数据,揭示新的科学门径,并加快改进。它还将在工程范围优化瞎想,提高遵循,推动更具改进性的时刻发展。
14.我我方也在日常责任中使用AI器用,来提高遵循和创造力。我认为,AI不仅能够匡助我们处理复杂的数据和决策任务,还能升迁我们的创酷爱维和责任遵循,成为每个东说念主责任中不可或缺的一部分。
以下是采访翰墨实录全文,由AI翻译:
主合手东说念主:Welcome back, Johnson, 30 years in to Nvidia and looking 10 years out, what are the big bets you think are still to make? Is it all about scale up from here? Are we running into limitations in terms of how we can squeeze more compute memory out of the architectures we have? What are you focused on? Well.
嗨,Johnson,接待转头!你在英伟达责任了30年,瞻望畴昔10年,你认为还有哪些迫切的投资契机?是不是说我们只需要连续扩大范围?我们在现存架构中是否会碰到罢了,无法再挤出更多的计较内存?你刻下关注的重心是什么?
黄仁勋:If we take a step back and think about what we‘ve done, we went from coding to machine learning, from writing software tools to creating AIs and all of that running on CPUs that was designed for human coding to now running on GPUs designed for AI coding, basically machine learning. And so the world has changed the way we do computing the whole stack has changed. And as a result, the scale of the problems we could address has changed a lot because we could, if you could paralyze your software on one GPU, you’ve set the foundations to paralyze across a whole cluster or maybe across multiple clusters or multiple data centers. And so I think we‘ve set ourselves up to be able to scale computing at a level and develop software at a level that nobody’s ever imagined before. And so we‘re at the beginning that over the next 10 years, our hope is that we could double or triple performance every year at scale, not at chip, at scale. And to be able to therefore drive the cost down by a factor of 2 or 3, drive the energy down by a factor of 2,3 every single year. When you do that every single year, when you double or triple every year in just a few years, it adds up. So it compounds really aggressively. And so I wouldn’t be surprised if, you know, the way people think about Moore‘s Law, which is 2 x every couple of years, you know, we’re gonna be on some kind of a hyper Moore‘s Law curve. And I fully hope that we continue to do that. Well, what.
以前我们编程都是靠我方写代码,当今我们运行让机器我方学习,我方写代码。以前我们用的那种电脑芯片(CPU)是给东说念主写代码用的,当今我们用的电脑芯片(GPU)是给机器学惯用的。因为这些变化,我们当今处理问题的方式和以前完全不一样了。打个比喻,如果你能让一个机器学习身手在一个GPU上运行,那你就不错让它在所有电脑群里,以致在好多电脑群或者数据中心里运行。这意味着我们当今能处理的问题比以前大多了。是以,我们确信我方仍是建立了能够大范围扩展计较才智和开发软件的基础,这个范围是以前没东说念主想象过的。
我们但愿在畴昔10年里,每年都能让计较才智翻两倍或者三倍,不是单个芯片的才智,而是举座的才智。这样的话,我们就能每年把计较老本责怪两倍或三倍,把能耗也责怪两倍或三倍。这种增长如果每年都能罢了,那么几年下来,这个增长会相等惊东说念主。因此,我认为畴昔的计较将会超越传统的“摩尔定律”(即每两年计较才智翻倍),可能会走上一条更快的增长弧线,我也相等但愿能够连续沿着这个方上前进。
主合手东说念主:Do you think is the driver of making that happen even faster than Morse law? Cuz I know morezo was sort of self reflexive, right? It was something that he said and then people kind of implemented it to me to happen.
你认为是什么身分推动了计较才智增长速率额外摩尔定律的?因为我知说念,摩尔定律自己便是一种“自我罢了”的门径,对吧?也便是说,摩尔定律自己是摩尔提倡的,然后群众就按照这个门径去作念,罢了它就罢清醒。
黄仁勋:Yep, too. Fundamental technical pillars. One of them was Denard scaling and the other one was Carver Mead‘s VLSI scaling. And both of those techniques were rigorous techniques, but those techniques have really run out of steam. And, and so now we need a new way of doing scaling. You know, obviously the new way of doing scaling are all kinds of things associated with co design. Unless you can modify or change the algorithm to reflect the architecture of the system or change and then change the system to reflect the architecture of the new software and go back and forth. Unless you can control both sides of it, you have no hope. But if you can control both sides of it, you can do things like
move from FP64 to FP32 to BF16 to FPA to, you know, FP4 to who knows what, right? And so, and so I think that code design is a very big part of that. The second part of it, we call it full stack. The second part of it is data center scale. You know, unless you could treat the network as a compute fabric and push a lot of the work into the network, push a lot of the work into the fabric. And as a result, you‘re compressing, you know, doing compressing at very large scales. And so that’s the reason why we bought Melanox and started fusing infinite and MV Link in such an aggressive way.
畴昔推动时刻向上的两个要害时刻柱子是Denard缩放(Denard Scaling)和Carver Mead的VLSI缩放。但是这两种方法当今都不太管用了,我们需要新的方法来变得更快。
新方式便是“协同瞎想”(co-design),也便是软件和硬件必须同期商量和优化。具体来说,如果你不可修改或转念算法,使其与系统的架构匹配,或者不可窜改系统架构,以适合新软件的需求,那么就莫得但愿。但如果你能同期限度软件和硬件,你就能作念好多新的事情,比如:从高精度的FP64转到低精度的FP32,再到BF16、FPA、以致FP4等更低精度的计较。
这便是为什么“协同瞎想”这样迫切的原因。另外,另一个迫切的部分是全栈瞎想。这意味着,你不仅要商量硬件,还要商量数据中心级别的范围。比如,必须把集合动作一个计较平台来使用,把多量的计较任务推到集合里,应用集合和硬件进行大范围压缩运算。
因此,我们收购了Mellanox,并运行相等积极地推动InfiniBand和NVLink这类高速衔接时刻,来扶持这种全新的大范围计较架构。
And now look where MV Link is gonna go. You know, the compute fabric is going to, I scale out what appears to be one incredible processor called a GPU. Now we get hundreds of GPUs that are gonna be working together.And now look where MV Link is gonna go. You know, the compute fabric is going to, I scale out what appears to be one incredible processor called a GPU. Now we get hundreds of GPUs that are gonna be working together.You know, most of these computing challenges that we‘re dealing with now, one of the most exciting ones, of course, is inference time scaling, has to do with essentially generating tokens at incredibly low latency because you’re self reflecting, as you just mentioned. I mean, you‘re gonna be doing tree surge, you’re gonna be doing chain of thought, you‘re gonna be doing probably some amount of simulation in your head. You’re gonna be reflecting on your own answers. Well, you‘re gonna be prompting yourself and generating text to your in, you know, silently and still respond hopefully in a second. Well, the only way to do that is if your latency is extremely low.Meanwhile, the data center is still about producing high throughput tokens because you know, you still wanna keep cost down, you wanna keep the throughput high, you wanna, right, you know, and generate a return. And so these two fundamental things about a factory, low latency and high throughput, they’re at odds with each other. And so in order for us to create something that is really great in both, we have to go invent something new, and Envy Link is really our way of doing that.We now you have a virtual GPU that has incredible amount of flops because you need it for context. You need a huge amount of memory, working memory, and still have incredible bandwidth for token generation all of the same time.
当今看NVLink(英伟达的高速衔接时刻)将走向那处,畴昔的计较架构将变得相等刚劲。你不错把它想象成一个超等刚劲的处理器,便是GPU(图形处理单位)。而当今,英伟达的方针是把数百个GPU集成到通盘,协同责任,造成一个深广的计较平台。
当今我们面对的计较挑战中,有一个相等令东说念主振作的问题便是推理时间的缩小。极度是在生成文本时,需要相等低的延长。因为就像你刚才提到的,我们的念念维其实是一种自我反念念的经过:你可能在脑海中进行“树形搜索”(tree search)、念念考链条(chain of thought),以致可能会进行某种模拟,讲求我方的谜底。你会我方给我方发问,并生成谜底,在大脑里“缄默地”念念考,然后但愿能在几秒钟内回话出来。
为了作念到这少许,计较的延长必须相等低,因为你不可能等太久才能得到罢了。
但与此同期,数据中心的任务是产生多量的高浑沌量的“token”(记号)。你需要限度老本,保合手高浑沌量,况且确保能够获取得报。是以,低延长和高浑沌量是两个互相矛盾的方针:低延长要求快速反应,而高浑沌量则需要处理更多的数据。这两者之间存在冲突。
为了同期作念到这两点,必须创造一些全新的时刻,而NVLink便是我们处置这个问题的方法之一。通过NVLink,英伟达但愿能够在确保高浑沌量的同期,也能提供低延长,从而处置这一计较上的矛盾,升迁举座性能。
当今我们有了编造GPU,它的计较才智相等刚劲,因为我们需要这样强的计较才智来处理荆棘文。也便是说,当我们在处理一些任务时,需要相等大的内存(极度是责任内存),同期还要有极高的带宽来生成token(即文本或数据记号)。
主合手东说念主:Building the models, actually also optimizing things pretty dramatically like David and my team pull data where over the last 18 months or so, the cost of 1 million tokens going into a GPT four equivalent model is basically dropped 240 x. Yeah, and so there‘s just massive optimization and compression happening on that side as.
构建模子的经过其实也包括了好多优化责任,比如David和他的团队,通过畴昔18个月的接力,得胜地将每百万个token的老本(用于GPT-4类模子的老本)责怪了240倍。
黄仁勋:Well. Just in our layer, just on the layer that we work on. You know, one of the things that we care a lot about, of course, is the ecosystem of our stack and the productivity of our software. You know, people forget that because you have Kuda Foundation and that‘s a solid foundation. Everything above it can change. If everything, if the foundation’s changing underneath you, it‘s hard to build a building on top. It’s hard to create anything and interesting on top. And so could have made it possible for us to iterate so quickly just in the last year. And then we just went back and benchmarked when Lama first came out, we‘ve improved the performance of Hopper by a factor of five without the algorithm, without the layer on top ever changing. Now, well, a factor of five in one year is impossible using traditional computing approaches. But it’s already computing and using this way of code design, we‘re able to explain all kinds of new things.
在我们的责任范围里,有一件相等迫切的事情,便是时刻栈的生态系统和软件的坐褥力。我们极度疼爱的是Kuda Foundation这个基础平台,它是非常褂讪和坚实的。因为如果基础平台无间变化,想要在上头构建出一个系统或者应用就相等贫瘠,根底无法在不褂讪的基础上创造出道理的东西。是以,Kuda Foundation的褂讪性让我们能够相等快速地进行迭代和改进,尤其是在畴昔一年里。
然后,我们还作念了一个对比测试:当Lama初次推出时,我们通过优化Hopper(一种计较平台或架构),在不窜改算法和不窜改表层架构的情况下,升迁了性能5倍。而且这种5倍的升迁,在传统的计较方法下是险些不可能罢了的。但通过协同瞎想这种新的方法,我们能够在现存的基础上无间改进息争释更多新的时刻可能性。
主合手东说念主:How much are, you know, your biggest customers thinking about the interchangeability of their infrastructure between large scale training and inference?
你的那些最大客户有多暖热他们在大范围磨练和推理之间基础设施的互换性?
黄仁勋:Well, you know, infrastructure is disaggregated these days. Sam was just telling me that he had decommissioned Volta just recently. They have pascals, they have amperes, all different configurations of blackwall coming. Some of it is optimized for air cool, some of it‘s optimized liquid cool. Your services are gonna have to take advantage of all of this. The advantage that Nvidia has, of course, is that the infrastructure that you built today for training will just be wonderful for inference tomorrow. And most of Chat GBT, I believe, are inferenced on the same type of systems that we’re trained on just recently. And so you can train on, you can inference on it. And so you‘re leaving a trail of infrastructure that you know is going to be incredibly good at inference, and you have complete confidence that you can then take that return on it, on the investment that you’ve had and put it into a new infrastructure to go scale with, you know you‘re gonna leave behind something of use and you know that Nvidia and the rest of the ecosystem are gonna be working on improving the algorithm so that the rest of your infrastructure improves by a factor of five, you know, in just a year. And so that motion will never change.
当今的基础设施不像以前那样是一成不变的了。比如Sam刚告诉我,他们最近淘汰了Volta型号的开导。他们有Pascal型号的,有Ampere型号的,还有好多不同成就的Blackwall型号行将到来。有些开导是优化了空气冷却的,有些则是优化了液体冷却的。你们的服务需要能够应用所有这些不同的开导。
英伟达的上风在于,你今天为磨练搭建的基础设施,将来会相等得当用于推理。我确信大多数的Chat GBT(可能是指大型语言模子)都是在最近磨练过的疏导类型的系统上进行推理的。是以你不错在这个系统上磨练,也不错在这个系统上进行推理。这样,你就留住了一条基础设施的轨迹,你知说念这些基础设施将来会相等得当进行推理,你完全有信心不错把之前投资的通告,参加到新的基础设施中去,扩大范围。你知说念你会留住一些有用的东西,而且你知说念英伟达和所有生态系统都在接力转换算法,这样你的其他基础设施在只是一年内就能提高五倍的遵循。是以这种趋势是不会变的。
And so the way that people will think about the infrastructures, yeah, even though I built it for training today, it‘s gotta be great for training. We know it’s gonna be great for inference. Inference is gonna be multi scale. 谈话东说念主 2 08:53 I mean, you‘re gonna take, first of all, in order to, the still smaller models could have a larger model that’s still from and so you‘re still gonna create these incredible a frontier models. They’re gonna be used for, of course, the groundbreaking work. You‘re gonna use it for synthetic data generation. You’re gonna use the models, the big models that teach smaller models and distill down to smaller models. And so there‘s a whole bunch of different things you can do, but in the end, you’re gonna have giant models all the way down to little tiny models. The little tiny models are gonna be quite effective, you know, not as generalizable, but quite effective. And so, you know, they‘re gonna perform very specific stunts incredibly well that one task. And we’re gonna see superhuman task in one little tiny domain from a little tiny model. Maybe you know, it‘s not a small language model, but you know, tiny language model, TLMs are, you know, whatever. Yeah, so I think we’re gonna see all kinds of sizes and we hope isn‘t right, just kind of like softwares today.
东说念主们看待基础设施的方式在变,就像我当今建的这个设施固然是为了磨练用的,但它也必须很得当磨练。我们知说念它将来也会相等得当作念推理。推接待有好多不同的范围。
我是说,你会有各式不同大小的模子。小模子不错从大模子那里学习,是以你如故会创造一些前沿的大模子。这些大模子会用来作念独创性的责任,用来生成合成数据,用来教小模子,然后把学问蒸馏给小模子。是以你不错作念的事情有好多,但临了你会有从巨大的模子到相等小的模子。这些小模子将会相等有用,固然它们不可通用,但在特定任务上会相等有用。它们会在某个特定任务上推崇得相等好,我们将会看到在某个小小的范围里,小模子能完成超乎东说念主类的任务。也许它不是一个微型的语言模子,但你知说念,便是微型语言模子,TLMs,归正便是雷同的东西。是以我以为我们会看到各式大小的模子,就像当今的软件一样。
Yeah, I think in a lot of ways, artificial intelligence allows us to break new ground in how easy it is to create new applications. But everything about computing has largely remained the same. For example, the cost of maintaining software is extremely expensive. And once you build it, you would like it to run on a large of an install base as possible. You would like not to write the same software twice. I mean, you know, a lot of people still feel the same way. You like to take your engineering and move them forward. And so to the extent that, to the extent that the architecture allows you, on one hand, create software today that runs even better tomorrow with new hardware that‘s great or software that you create tomorrow, AI that you create tomorrow runs on a large install base. You think that’s great. That way of thinking about software is not gonna.
我以为在好多方面,东说念主工智能让我们能够更容易地创造新的应用身手。但是在计较方面,大部分事情如故老样式。比如说,维护软件的老本相等高。一朝你建好了软件,你但愿它能在尽可能多的开导上运行。你不想重叠写相似的软件。我的酷爱是,好多东说念主如故这样想的。你心爱把你的工程推上前进。是以,如果架构允许你,一方面,今天创建的软件未来在新硬件上能运行得更好,那就太好了;或者你未来创建的软件,后天创建的东说念主工智能能在好多开导上运行。你认为那很棒。这种商量软件的方式是不会变的。
主合手东说念主:Change. And video has moved into larger and larger, let‘s say, like a unit of support for customers. I think about it going from single chip to, you know, server to rack and real 72. How do you think about that progression? Like what’s next? Like should Nvidia do you full data center? But
跟着时刻的发展,英伟达的产物仍是不单是是单个的芯片了,而是扩展到了扶持所有数据中心的范围。你怎么看待这种发展?接下来会是什么?比如,英伟达是不是应该作念所有数据中心?
黄仁勋:In fact, we build full data centers the way that we build everything. Unless you‘re building, if you’re developing software, you need the computer in its full manifestation. We don‘t build Powerpoint slides and ship the chips and we build a whole data center. And until we get the whole data center built up, how do you know the software works until you get the whole data center built up, how do you know your, you know, your fabric works and all the things that you expected the efficiencies to be, how do you know it’s gonna really work at scale? And that‘s the reason why it’s not unusual to see somebody‘s actual performance be dramatically lower than their peak performance, as shown in Powerpoint slides, and it is, computing is just not used to, is not what it used to be. You know, I say that the new unit of computing is the data center. That’s to us. So that‘s what you have to deliver. That’s what we build.Now we build a whole thing like that. And then we, for every single thing that every combination, air cold, x 86, liquid cold, Grace, Ethernet, infinite band, MV link, no NV link, you know what I‘m saying? We build every single configuration. We have five supercomputers in our company today. Next year, we’re gonna build easily five more. So if you‘re serious about software, you build your own computers if you’re serious about software, then you‘re gonna build your whole computer. And we build it all at scale.
本色上,我们建造完竣的数据中心就像我们建造其他所有东西一样。如果你在开发软件,你需要电脑的完竣形态来测试。我们不单是作念PPT幻灯片然后发货芯片,我们建造所有数据中心。唯有当我们把所有数据中心搭建起来后,你才能知说念软件是否平素责任,你的集合布线是否有用,所有你欲望的遵循是否都能达到,你才知说念它是否真的能在大范围上运行。这便是为什么东说念主们的本色性能频频远低于PPT幻灯片上展示的峰值性能,计较仍是不再是畴昔的样式了。我说当今的计较单位是数据中心,对我们来说便是这样。这便是你必须委用的东西,亦然我们建造的东西。
我们当今就这样建造所有系统。然后我们为每一种可能的组合建造:空气冷却、x86架构、液体冷却、Grace芯片、以太网、无尽带宽、MVLink,莫得NVLink,你懂我的酷爱吗?我们建造每一种成就。我们公司当今有五台超等计较机,来岁我们骄傲就能再建造五台。是以,如果你对软件是稳健的,你就会我方建造计较机,如果你对软件是稳健的,你就会建造所有计较机。我们都是大范围地建造。
This is the part that is really interesting. We build it at scale and we build it very vertically integrate. We optimize it full stack, and then we disagree everything and we sell lemon parts. That‘s the part that is completely, utterly remarkable about what we do. The complexity of that is just insane. And the reason for that is we want to be able to graft our infrastructure into GCP, AWS, Azure, OCI. All of their control planes, security planes are all different and all of the way they think about their cluster sizing all different. And, but yet we make it possible for them to all accommodate Nvidia’s architecture. So that could, it could be everywhere. That‘s really in the end the singular thought, you know, that we would like to have a computing platform that developers could use that’s largely consistent, modular, you know, 10% here and there because people‘s infrastructure are slightly optimized differently and modular 10% here and there, but everything they build will run everywhere. This is kind of the one of the principles of software that should never be given up. And it, and we protected quite dearly. Yeah, it makes it possible for our software engineers to build ones run everywhere. And that’s because we recognize that the investment of software is the most expensive investment, and it‘s easy to test.
这部分真的很道理。我们不仅大范围建造,而且是垂直整合建造。我们从底层到顶层全程优化,然后我们把各个部分分开,单独卖。我们作念的事情复杂得让东说念主难以置信。为什么这样作念呢?因为我们想把我们的基础设施融入到GCP、AWS、Azure、OCI这些不同的云服务提供商中。我们的限度平台、安全平台都不一样,我们商量集群大小的方式也各不疏导。但是,我们如故想主见让他们都能适合英伟达的架构。这样,我们的架构就能无处不在。
最终,我们但愿有一个计较平台,开发者不错用它来构建软件,这个平台在大部分情况下是一致的,不错模块化地转念,可能这里那里有10%的不同,因为每个东说念主的基础设施都略有优化各别,但是无论在那处, 我们构建的东西都能运行。这是软件的一个原则,我们相等惊奇这少许。这使得我们的软件工程师不错构建出到处都能运行的软件。这是因为我们知道到,软件的投资是最腾贵的投资,而且它很容易测试。
Look at the size of the whole hardware industry and then look at the size of the world‘s industries. It’s $100 trillion on top of this one trillion dollar industry. And that tells you something.The software that you build, you have to, you know, you basically maintain for as long as you shall live. We‘ve never given up on piece of software. The reason why Kuda is used is because, you know, I called everybody. We will maintain this for as long as we shall live. And we’re serious now. We still maintain. I just saw a review the other day, Nvidia Shield, our Android TV. It‘s the best Android TV in the world. We shifted seven years ago. It is still the number one Android TV that people, you know, anybody who enjoys TV. And we just updated the software just this last week and people wrote a new story about it. G Force, we have 300 million gamers around the world. We’ve never stranded a single one of them. And so the fact that our architecture is compatible across all of these different areas makes it possible for us to do it. Otherwise, we would be sub, we would be, we would have, you know, we would have software teams that are hundred times the size of our company is today if not for this architectural compatibility. So we‘re very serious about that, and that translates to benefits the developers.
望望所有硬件行业的范围,再比比全宇宙所有行业的范围。硬件行业唯有一万亿好意思元,而全宇宙的行业加起来有一百万亿亿好意思元。这个对比告诉你,软件行业要比硬件行业大得多。
你们作念的软件,基本上要一直维护下去。我们从莫得废弃过任何一款软件。Kuda之是以被群众用,是因为我向所有东说念主承诺,我们会一直维护它,只须我们还在。我们当今如故很稳健的,我们还在维护它。我前几天还看到一篇驳倒,说我们的英伟达Shield,我们的安卓电视,是宇宙上最佳的安卓电视。我们在七年前推出的,它仍然是排行第一的安卓电视,任何心爱看电视的东说念主都爱它。我们上周才更新了软件,然后东说念主们就写了新的著述来驳倒它。我们的G Force,全宇宙有3亿玩家。我们从莫得放手过他们中的任何一个。我们的架构在所有这些不同范围都是兼容的,这使得我们能作念到这少许。如果不是因为我们的架构兼容性,不然我们今天的软件团队的范围会比当今公司大一百倍。是以我们相等疼爱这少许,这也给开发者带来了公道。
主合手东说念主:One impressive substantiation of that recently was how quickly brought up a cluster for X dot AI. Yeah, and if you want to check about that, cuz that was striking in terms of both the scale and the speed with what you did. That
最近有一个让东说念主印象深刻的例子是我们为X dot AI飞速搭建了一个集群。如果你想了解这件事,因为它在范围和速率上都让东说念主骇怪。我们很快就完成了这个任务。
黄仁勋:You know, a lot of that credit you gotta give to Elon. I think the, first of all, to decide to do something, select the site. I bring cooling to it. I power hum and then decide to build this hundred thousand GPU super cluster, which is, you know, the largest of its kind in one unit. And then working backwards, you know, we started planning together the date that he was gonna stand everything up. And the date that he was gonna stand everything up was determined, you know, quite, you know, a few months ago. And so all of the components, all the Oems, all the systems, all the software integration we did with their team, all the network simulation we simulate all the network configurations, we, we pre, I mean like we prestaged everything as a digital twin. We, we pres, we prestaged all of his supply chain. We prestaged all of the wiring of the networking. We even set up a small version of it. Kind of a, you know, just a first instance of it. You know, ground truth, if you reference 0, you know, system 0 before everything else showed up. So by the time that everything showed up, everything was staged, all the practicing was done, all the simulations were done.
这里得给埃隆·马斯克好多功劳。最初,他决定要作念这件事,选了方位,处置了冷却和供电问题,然后决定建造这个十万GPU的超等计较机群,这是迄今为止这种类型中最大的一个。然后,我们运行倒推,便是说,我们几个月前就通盘规划了他要让一切运行起来的日历。是以,所有的组件、所有的原始开导制造商、所有的系统、所有的软件集成,我们都是和他们的团队通盘作念的,所有的集合成就我们都模拟了一遍,我们事先准备,就像数字孪生一样,我们事先准备了所有的供应链,所有的集合布线。我们以致搭建了一个小版块,就像是第一个实例,你懂的,便是所有东西到位之前的基准,你参考的0号系统。是以,当所有东西都到位的时候,一切都仍是安排好了,所有的训练都作念收场,所有的模拟也都完成了。
And then, you know, the massive integration, even then the massive integration was a Monument of, you know, gargantuan teams of humanity crawling over each other, wiring everything up 247. And within a few weeks, the clusters were out. I mean, it‘s, it’s really, yeah, it‘s really a testament to his willpower and how he’s able to think through mechanical things, electrical things and overcome what is apparently, you know, extraordinary obstacles. I mean, what was done there is the first time that a computer of that large scale has ever been done at that speed. Unless our two teams are working from a networking team to compute team to software team to training team to, you know, and the infrastructure team, the people that the electrical engineers today, you know, to the software engineers all working together. Yeah, it‘s really quite a fit to watch. Was.
然后,你知说念,大范围的集成责任,即使这个集成责任自己亦然个巨大的工程,需要多量的团队成员像蚂蚁一样忙活责任,险些是全天候无间地接线和竖立。几周之内,这些计较机群就建成了。这真的是对他意志力的解释,也骄傲了他如安在机械、电气方面念念考,并克服了赫然是非常巨大的缺乏。我的酷爱是,这关联词第一次在这样短的时间内建成如斯大范围的计较机系统。这需要我们的集合团队、计较团队、软件团队、磨练团队,以及基础设施团队,也便是那些电气工程师、软件工程师,所有东说念主通盘合作。这真的挺壮不雅的。这就像是一场大型的团队协调,每个东说念主都在接力确保一切顺利运行。
主合手东说念主:There a challenge that felt most likely to be blocking from an engineering perspective, active, just.
从工程角度来看,有莫得哪个挑战最可能成为绊脚石,便是说,有莫得哪个时刻难题最可能让所有技俩卡住,滚动不得?
黄仁勋:A tonnage of electronics that had to come together. I mean, it probably worth just to measure it. I mean, it‘s a, you know, it tons and tons of equipment. It’s just abnormal. You know, usually a supercomputer system like that, you plan it for a couple of years from the moment that the first systems come on, come delivered to the time that you‘ve probably submitted everything for some serious work. Don’t be surprised if it‘s a year, you know, I mean, I think that happens all the time. It’s not abnormal. Now we couldn‘t afford to do that. So we created, you know, a few years ago, there was an initiative in our company that’s called Data Center as a product. We don‘t sell it as a product, but we have to treat it like it’s a product. Everything about planning for it and then standing it up, optimizing it, tuning it, keep it operational, right? The goal is that it should be, you know, kind of like opening up your beautiful new iPhone and you open it up and everything just kind of works.
我们需要把多量的电子开导整合在通盘。我的酷爱是,这些开导的量多到值得去称一称。罕有吨又数吨的开导,这太对抗素了。频频像这样的超等计较机系统,从第一个系统运行委用,到你把所有东西都准备好进行一些严肃的责任,你频频需要磋商几年时间。如果这个经过需要一年,你要知说念,这是常有的事,并不奇怪。
但当今我们莫得时间去这样作念。是以几年前,我们公司里有一个叫作念“数据中心即产物”的规划。我们不把它动作产物来卖,但我们必须像对待产物一样对待它。从磋商到建立,再到优化、转念、保合手运行,所有的一切都是为了确保它能够像掀开一部簇新的iPhone一样,一掀开,一切都能平素责任。我们的方针便是这样。
Now, of course, it‘s a miracle of technology making it that, like that, but we now have the skills to do that. And so if you’re interested in a data center and just have to give me a space and some power, some cooling, you know, and we‘ll help you set it up within, call it, 30 days. I mean, it’s pretty extraordinary.
天然了,能这样快就把数据中心建好,这简直便是科技的古迹。但当今我们仍是有了这样的时刻才智。是以如果你想要建一个数据中心,只需要给我一个方位,提供一些电力和制冷开导,我们就能在差未几30天内帮你把一切都搭建好。我的酷爱是,这真的相等了不得。
主合手东说念主:That‘s wild. If you think, if you look ahead to 200,000,500,000, a million in a super cluster, whatever you call it. At that point, what do you think is the biggest blocker? Capital energy supply in one area?
那真的犀利。如果你想想,淌若将来有个超等大的计较机集群,内部有个二十万、五十万、以致一百万的计较机,不管你叫它什么。到阿谁时候,你以为最大的难题会是什么呢?是资金问题、动力供应问题,如故别的什么?
黄仁勋:Everything. Nothing about what you, just the scales that you talked about, though, nothing is normal.
你说的那些事情,不管是哪个方面,只须波及到你提到的那些巨大范围,那就莫得一件事情是平素的。
主合手东说念主:But nothing is impossible. Nothing.
但是,也没什么事是完全不可能的。啥事都有可能。
黄仁勋:Is, yeah, no laws of physics limits, but everything is gonna be hard. And of course, you know, I, is it worth it? Like you can‘t believe, you know, to get to something that we would recognize as a computer that so easily and so able to do what we ask it to do, what, you know, otherwise general intelligence of some kind and even, you know, even if we could argue about is it really general intelligence, just getting close to it is going to be a miracle. We know that. And so I think the, there are five or six endeavors to try to get there. Right? I think, of course, OpenAI and anthropic and X and, you know, of course, Google and meta and Microsoft and you know, there, this frontier, the next couple of clicks that mountain are just so vital. Who doesn’t wanna be the first on that mountain. I think that the prize for reinventing intelligence altogether. Right. It‘s just, it’s too consequential not to attempt it. And so I think there are no laws of physics. Everything is gonna be hard.
照实,莫得物理定律说我们作念不到,但每件事情都会相等难。你也知说念,这值得吗?你可能以为难以置信,我们要达到的那种电脑,能够消弱地作念我们让它作念的事情,也便是某种通用智能,哪怕我们能争论它是否真的是通用智能,接近它都将会是一个古迹。我们知说念这很难。是以我认为,有五六个团队正在尝试达到这个方针。对吧?比如说,OpenAI、Anthropic、X,还有谷歌、Meta和微软等等,他们都在接力攀高这个前沿科技的山岭。谁不想成为第一个登顶的东说念主呢?我认为,再行发理智能的奖励是如斯之大,它的影响太大了,我们不可不去尝试。是以,固然物理定律上莫得罢了,但每件事都会很难。
主合手东说念主:A year ago when we spoke together, you talked about, we asked like what applications you got most excited about that Nvidia would serve next in AI and otherwise, and you talked about how you led to, your most extreme customers sort of lead you there. Yeah, and about some of the scientific applications. So I think that‘s become like much more mainstream of you over the last year. Is it still like science and AI’s application of science that most excites you?
一年前我们聊天时,我问你,你对英伟达接下来在AI和其他范围能服务的哪些应用最振作,你谈到了你的一些最顶点的客户某种进度上团结了你。是的,还关系于一些科学应用的接洽。是以我以为畴昔一年里,这些科学和AI的应用变得更主流了。当今,是不是仍然是科学以及AI在科学范围的应用让你最振作?
黄仁勋:I love the fact that we have digital, we have AI chip designers here in video. Yeah, I love that. We have AI software engineers. How.
我就直说了,我们当今罕有字版的,也便是用东说念主工智能来瞎想芯片的瞎想师,就在视频里。是的,我心爱这个。我们还有AI软件工程师。
主合手东说念主:Effective our AI chip designers today? Super.
我们今天用东说念主工智能来瞎想芯片的罢了怎么样?相等好。
黄仁勋:Good. We can‘t, we couldn’t build Hopper without it. And the reason for that is because they could explore a much larger space than we can and because they have infinite time. They‘re running on a supercomputer. We have so little time using human engineers that we don’t explore as much of the space as we should, and we also can explore commentary. I can‘t explore my space while including your exploration and your exploration. And so, you know, our chips are so large, it’s not like it‘s designed as one chip. It’s designed almost like 1,000 ships and we have to ex, we have to optimize each one of them. Kind of an isolation. You really wanna optimize a lot of them together and, you know, cross module code design and optimize across much larger space. But obviously we‘re gonna be able to find fine, you know, local maximums that are hidden behind local minimum somewhere. And so clearly we can find better answers. You can’t do that without AI. Engineers just simply can‘t do it. We just don’t have enough time.
我们的AI芯片瞎想师真的很犀利。如果莫得它们,我们根底造不出Hopper这款芯片。因为它们能探索的范围比我们东说念主类广得多,而且它们好像有取之不尽的时间。它们在超等计较机上运行,而我们东说念主类工程师的时间有限,探索不了那么大的范围。而且,我们也不可同期探索所有的可能,我探索我的范围的时候,就不可同期探索你的范围。
我们的芯片相等大,不像是瞎想一个单独的芯片,更像是瞎想1000个芯片,每个都需要优化。就像是一个个孤独的小岛。但我们其实很想把它们放在通盘优化,跨模块协同瞎想,在所有更大的空间里优化。赫然,我们能找到更好的处置决议,那些笼罩在某个边际里的最佳的遴荐。莫得AI我们作念不到这少许。工程师们便是时间不够,作念不到。
主合手东说念主:One other thing has changed since we last spoke collectively, and I looked it up at the time in videos, market cap was about 500 billion. It‘s now over 3 trillion. So the last 18 months, you’ve added two and a half trillion plus of market cap, which effectively is $100 billion plus a month or two and a half snowflakes or, you know, a stripe plus a little bit, or however you wanna think about.A country or two. Obviously, a lot of things are stayed consistent in terms of focus on what you‘re building and etc. And you know, walking through here earlier today, I felt the buzz like when I was at Google 15 years ago was kind of you felt the energy of the company and the vibe of excitement. What has changed during that period, if anything? Or how, what is different in terms of either how Nvidia functions or how you think about the world or the size of bets you can take or.
自我们前次通盘聊天以来,有一件事变了,我查了下,那时英伟达的市值或者是5000亿好意思元。当今额外了3万亿好意思元。是以在畴昔18个月里,你们加多了两万五千亿好意思元以上的市值,这额外于每个月加多了1000亿好意思元,或者说加多了两个半的Snowflake公司或者一个Stripe公司多少许的市值,无论你怎么想。
这额外于加多了一两个国度的市值。赫然,尽管市值增长了这样多,你们在建造的东西和专注的范围上如故保合手了一致性。你知说念,今天我在这里走了一圈,我感受到了一种活力,就像15年前我在谷歌时感受到的那样,你能嗅觉到公司的能量和振作的氛围。在这段时间里,有什么变化了吗?或者,英伟达的运作方式、你对宇宙的看法、你能承担的风险大小等方面有什么不同了吗?
黄仁勋:Well, our company can‘t change as fast as a stock price. Let’s just be clear about. So in a lot of ways, we haven‘t changed that much. I think the thing to do is to take a step back and ask ourselves, what are we doing? I think that’s really the big, you know, the big observation, realization, awakening for companies and countries is what‘s actually happening. I think what we’re talking about earlier, I‘m from our industry perspective, we reinvented computing. Now it hasn’t been reinvented for 60 years. That‘s how big of a deal it is that we’ve driven down the marginal cost of computing, down probably by a million x in the last 10 years to the point that we just, hey, let‘s just let the computer go exhaustively write the software. That’s the big realization. 谈话东说念主 2 24:00 And that in a lot of ways, I was kind of, we were kind of saying the same thing about chip design. We would love for the computer to go discover something about our chips that we otherwise could have done ourselves, explore our chips and optimize it in a way that we couldn‘t do ourselves, right, in the way that we would love for digital biology or, you know, any other field of science.
我们公司的变化速率可莫得股价变化那么快。是以这样说吧,我们在好多方面并莫得太大变化。我认为迫切的是要退一步来问问我们我方,我们到底在作念什么。这真的是对公司和国度来说一个很大的不雅察、知道和觉悟,那便是真实发生的事情。
就像我们之前接洽的,从我们行业的角度来看,我们再行发明了计较。这关联词60年来都莫得发生过的事情。我们把计较的边际老本责怪了,可能在畴昔10年里责怪了一百万分之一,以至于我们当今不错让计较机去详备地编写软件。这是一个要紧的知道。
在好多方面,我们对芯片瞎想亦然这样说的。我们但愿计较机能我方去发现我们芯片的一些东西,这些东西我们蓝本不错我方作念,但计较机不错探索我们的芯片并以我们我方作念不到的方式进行优化,就像我们但愿在数字生物学或其他科学范围那样。
And so I think people are starting to realize when we reinvented computing, but what does that mean even, and as we, all of a sudden, we created this thing called intelligence and what happened to computing? Well, we went from data centers are multi tenant stores of files. These new data centers we‘re creating are not data centers. They don’t, they‘re not multi tenant. They tend to be single tenant. They’re not storing any of our files. They‘re just, they’re producing something. They‘re producing tokens. And these tokens are reconstituted into what appears to be intelligence. Isn’t that right? And intelligence of all different kinds. You know, it could be articulation of robotic motion. It could be sequences of amino acids. It could be, you know, chemical chains. It could be all kinds of interesting things, right? So what are we really doing? We‘ve created a new instrument, a new machinery that in a lot of ways is that the noun of the adjective generative AI. You know, instead of generative AI, you know, it’s, it‘s an AI factory. It’s a factory that generates AI. And we‘re doing that at extremely large scale. And what people are starting to realize is, you know, maybe this is a new industry. It generates tokens, it generates numbers, but these numbers constitute in a way that is fairly valuable and what industry would benefit from it.
是以我以为东说念主们运行坚强到,当我们再行发明计较时,这到底意味着什么。片刻间,我们创造了这个叫作念智能的东西,计较发生了什么变化?嗯,我们以前把数据中心看作是多佃农存储文献的方位。我们当今创建的这些新数据中心,其实仍是不是传统真理上的数据中心了。它们往往是单一佃农的,它们不存储我们的文献,它们只是在坐褥一些东西。它们在坐褥数据令牌。然后这些数据令牌再行组合成看起来像智能的东西。对吧?而且智能有各式万般的体式。可能是机器东说念主动作的抒发,可能是氨基酸序列,可能是化学物资链,可能是各式道理的事情,对吧?是以我们到底在作念什么?我们创造了一种新的器用,一种新的机械,从好多方面来说,它便是生成性东说念主工智能的名词体式。你知说念,不是生成性东说念主工智能,而是东说念主工智能工场。它是一个坐褥东说念主工智能的工场。我们正在相等大范围地作念这件事。东说念主们运行坚强到,这可能是一个新行业。它生成数据令牌,它生成数字,但这些数字以一种额外有价值的方式组成,哪些行业会从中受益。
Then you take a step back and you ask yourself again, you know, what‘s going on? Nvidia on the one hand, we reinvent a computing as we know it. And so there’s $1 trillion of infrastructure that needs to be modernized. That‘s just one layer of it. The big layer of it is that there’s, this instrument that we‘re building is not just for data centers, which we were modernizing, but you’re using it for producing some new commodity. And how big can this new commodity industry be? Hard to say, but it‘s probably worth trillions. 谈话东说念主 2 26:18 And so that I think is kind of the viewers to take a step back. You know, we don’t build computers anymore. We build factories. And every country is gonna need it, every company‘s gonna need it, you know, give me an example of a company who or industry as us, you know what, we don’t need to produce intelligence. We got plenty of it. And so that‘s the big idea. I think, you know, and that’s kind of an abstracted industrial view. And, you know, someday people realize that in a lot of ways, the semiconductor industry wasn‘t about building chips, it was building, it was about building the foundational fabric for society. And then all of a sudden, there we go. I get it. You know, this is a big deal. Isn’t not just about chips.
然后你退一步,再次问我方,到底发生了什么?Nvidia一方面,我们再行发明了我们所知说念的计较。是以有一万亿好意思元的基础设施需要当代化。这只是其中一层。更大的一层是,我们正在建造的这个器用不单是是为了数据中心,我们正在当代化数据中心,而是你用它来坐褥一些新的商品。这个新商品行业能有多大?很难说,但可能价值数万亿好意思元。
是以我认为这是不雅众需要退一步的方位。你知说念,我们不再制造电脑了。我们制造工场。每个国度都会需要它,每个公司都会需要它,给我一个不需要坐褥智能的公司或行业的例子,你知说念,我们有好多智能。是以这便是这个大主意。我认为,你知说念,这是一种笼统的工业不雅点。然后,有一天东说念主们坚强到,在好多方面,半导体行业不是对于制造芯片,它是对于为社会建立基础结构。然后片刻间,我们分解了。这不单是是对于芯片的大事。
主合手东说念主:How do you think about embodiment now?
你当今怎么看待“体现”或者“具体化”这个意见?便是说,你怎么商量把智能或者东说念主工智能真实应用到本色的物理宇宙中,比如机器东说念主或者其他实体开导上?
黄仁勋:Well, the thing I‘m super excited about is in a lot of ways, we’ve, we‘re close to artificial general intelligence, but we’re also close to artificial general robotics. Tokens are tokens. I mean, the question is, can you tokenize it? You know, of course, tokenis, tokenizing things is not easy, as you guys know. But if you‘re able to tokenize things, align it with large language models and other modalities, if I can generate a video that has Jensen reaching out to pick up the coffee cup, why can’t I prompt a robot to generate the token, still pick up the rule, you know? And so intuitively, you would think that the problem statement is rather similar for computer. And, and so I think that we‘re that close. That’s incredibly exciting.
我当今相等振作的少许是,我们在好多方面都将近罢了通用东说念主工智能了,而且我们也快罢了通用机器东说念主时刻了。数据令牌便是数据令牌。我的酷爱是,问题是,你能把它变成数据令牌吗?天然,把东西变成数据令牌并辞谢易,你们知说念这少许。但如果你能作念到这少许,把它和大型语言模子和其他方式对皆,如果我能生成一个视频,视频里有Jensen伸手去拿咖啡杯,为什么我不可教唆一个机器东说念主去生成数据令牌,本色上去提起阿谁轨则,你知说念吗?是以直不雅上,你会认为这个问题对计较机来说额皮毛似。是以我认为我们仍是很接近了。这相等令东说念主振作。
Now the, the two brown field robotic systems. Brown field means that you don‘t have to change the environment for is self driving cars. And with digital chauffeurs and body robots right between the cars and the human robot, we could literally bring robotics to the world without changing the world because we built a world for those two things. Probably not a coincidence that Elon spoke is then those two forms. So robotics because it is likely to have the larger potential scale. And and so I think that’s exciting. But the digital version of it, I is equally exciting. You know, we‘re talking about digital or AI employees. There’s no question we‘re gonna have AI employees of all kinds, and our outlook will be some biologics and some artificial intelligence, and we will prompt them in the same way. Isn’t that right? Mostly I prompt my employees, right? You know, provide them context, ask him to perform a mission. They go and recruit other team members, they come back and work going back and forth. How‘s that gonna be any different with digital and AI employees of all kinds? So we’re gonna have AI marketing people, AI chip designers, AI supply chain people, AIs, you know, and I‘m hoping that Nvidia is someday biologically bigger, but also from an artificial intelligence perspective, much bigger. That’s our future company. If.
当今有两种“棕色地带”机器东说念主系统。“棕色地带”意味着你不需要窜改环境,比如自动驾驶汽车。有了数字司机和机器东说念主助手在汽车和东说念主类机器东说念主之间,我们不错在不窜改宇宙的情况下把机器东说念主时刻带到宇宙上,因为我们为这两样东西建造了宇宙。埃隆·马斯克可能不是未必提到这两种体式的。是以机器东说念主时刻因为可能有更大的潜在范围而令东说念主振作。而数字版的机器东说念主也相似令东说念主振作。你知说念,我们议论的是数字或AI职工。毫无疑问,我们将领有各式AI职工,我们的远景将是一些生物和一些东说念主工智能,我们将以疏导的方式教唆他们。不是吗?大多数情况下,我教唆我的职工,对吧?给他们提供荆棘文,让他们蔓延任务。他们去招募其他团队成员,他们转头责任,往还责任。这和各式数字和AI职工有什么不同呢?是以我们将有AI营销东说念主员,AI芯片瞎想师,AI供应链东说念主员,AI,等等,我但愿英伟达有一天在生物学上更大,同期从东说念主工智能的角度来看,也更大。这是我们畴昔公司的样式。
主合手东说念主:We came back and talked to you year from now, what part of the company do you think would be most artificially intelligent?
如果我们一年后转头再和你聊聊,你以为公司里哪个部分会是最智能化的?
黄仁勋:I‘m hoping it should sign.
我但愿公司里最迫切的、最中枢的部分能罢了智能化。
主合手东说念主:Okay. And most.
好的,然后连续连络。
黄仁勋:Important part. And the read. That‘s right. Because it because I should start where it moves the needle most also where we can make the biggest impact most. You know, it’s such an insanely hard problem. I work with Sasina at synopsis and rude at cadence. I totally imagine them having synopsis chip designers that I can rent. And they know something about a particular module, their tool, and they train an AI to be incredibly good at it. And we‘ll just hire a whole bunch of them whenever we need, we’re in that phase of that chip design. You know, I might rent a million synopsis engineers to come and help me out and then go rent a million Cadence engineers to help me out. And that, what an exciting future for them that they have all these agents that sit on top of their tools platform, that use the tools platform and other, and collaborate with other platforms. And you‘ll do that for, you know, Christian will do that at SAP and Bill will do that as service.
我认为最迫切的部分应该是公司里最能产生影响的方位。他说,这个问题相等难,但他但愿从最能推动公司发展的方位运行智能化。他和Synopsys的Sasina和Cadence的Rude通盘责任,他想象着不错租用Synopsys的芯片瞎想师AI。这些AI对某个特定模块、器用相等了解,况且仍是被磨练得相等擅长这方面的责任。当他们需要进行芯片瞎想的某个阶段时,他们会租用一多量这样的AI瞎想师。比如,他可能会租用一百万个Synopsys工程师AI来维护,然后再租用一百万个Cadence工程师AI来维护。我认为,对于我们来说,有一个慷慨东说念主心的畴昔,因为我们有所有这些AI代理,它们位于我们器用平台的顶部,使用这些器用平台,况且与其他平台协调。SAP的Christian会这样作念,Bill会作为服务来作念这件事。
Now, you know, people say that these Saas platforms are gonna be disrupted. I actually think the opposite, that they‘re sitting on a gold mine, that they’re gonna be this flourishing of agents that are gonna be specialized in Salesforce, specialized in, you know, well, Salesforce, I think they call Lightning and SAP is about, and everybody‘s got their own language. Is that right? And we got Kuda and we’ve got open USD for Omniverse. And who‘s gonna create an AI agent? That’s awesome. At open USD, we‘re, you know, because nobody cares about it more than we do, right? And so I think in a lot of ways, these platforms are gonna be flourishing with agents and we’re gonna introduce them to each other and they‘re gonna collaborate and solve problems.
当今,有些东说念主说这些基于集合的软件服务平台(SaaS)将会被颠覆。但我本色上认为偶合相背,他们就像坐在金矿上一样,将会有一个专科化的智能代理(AI)的郁勃时期。这些智能代理将会特地针对Salesforce、SAP等平台进行优化。比如Salesforce有个叫作念Lightning的平台,每个平台都有我方的语言和特色。我们有Kuda,还有为Omniverse准备的怒放USD。谁会来创造这些AI代理呢?那将会是非常酷的事情。在怒放USD方面,我们会来作念,因为莫得东说念主比我们更暖热它,对吧?是以我认为在好多方面,这些平台将会因为这些智能代理而郁勃起来,我们会把它们互相先容,它们将会协调并处置问题。
主合手东说念主:You see a wealth of different people working in every domain in AI. What do you think is under notice or that people that you want more entrepreneurs or engineers or business people could work on?
你以为在东说念主工智能范围,有莫得什么被冷落的方位,或者你但愿更多的创业者、工程师或营业东说念主士能关注和参加责任的范围?
黄仁勋:Well, first of all, I think what is misunderstood, and I misunderstood, maybe it may be underestimated, is the, the under the water activity, under the surface activity of groundbreaking science, computer science to science and engineering that is being affected by AI and machinery. I think you just can‘t walk into a science department anywhere, theoretical math department anywhere, where AI and machine learning and the type of work that we’re talking about today is gonna transform tomorrow. If they are, if you take all of the engineers in the world, all of the scientists in the world and you say that the way they‘re working today is early indication of the future, because obviously it is. Then you’re gonna see a tidal wave of gender to AI, a tidal wave of AI, a tidal wave machine learning change everything that we do in some short period of time.
最初,我认为可能被诬陷或低估了的是,那些在水面下的、正在进行的、冲突性的科学、计较机科学以及科学与工程行动,这些行动正受到东说念主工智能和机械的影响。如果你走进任何一个科学系,任何一个表面数学系,你会发现今天的东说念主工智能和机器学习的责任将窜改未来。如果你把宇宙上所有的工程师、所有的科学家都看作是畴昔的早期迹象,因为赫然他们是,那么你就会看到一股涌向东说念主工智能的潮水,一鼓舞说念主工智能的潮水,一股机器学习窜改我们所作念的一切的潮水,这将在很短的时间内发生。
in some short period of time.ion. And to work with Alex and Elian and Hinton at at at in Toronto and Yan Lekun and of course, Andrew Ang here in Stanford. And, you know, I saw the early indications of it and we were fortunate to have extrapolated from what was observed to be detecting cats into a profound change in computer science and computing altogether. And that extrapolation was fortunate for us. And now, of course, we, we were so excited by, so inspired by it that we changed everything about how we did things. But that took how long? It took literally six years from observing that toy, Alex Net, which I think by today‘s standards will be considered a toy to superhuman levels of capabilities in object recognition. Well, that was only a few years. 谈话东说念主 2 33:40 Now what is happening right now, the groundswell in all of the fields of science, not one field of science left behind. I mean, just to be very clear. Okay, everything from quantum computing, the quantum chemistry, you know, every field of science is involved in the approaches that we’re talking about. If we give ourselves, and they‘ve been added for a couple to three years, if we give ourselves in a couple, two, three years, the world’s gonna change. There‘s not gonna be one paper, there’s not gonna be one breakthrough in science, one breakthrough in engineering, where generative AI isn‘t at the foundation of it. I’m fairly certain of it. And, and so I, I think, you know, there‘s a lot of questions about, you know, every so often I hear about whether this is a fad computer. You just gotta go back to first principles and observe what is actually happening.
就在很短的时间内,我们看到了科学范围的大波澜,莫得一个科学范围被落下。我的酷爱是,每一件事都相等清醒。从量子计较到量子化学,你知说念的,每个科学范围都波及到我们正在接洽的方法。如果我们给我方,比如说,两三年的时间,宇宙将会窜改。不会有一篇科学论文,不会有一项科学冲突,一项工程冲突,不是以生成性东说念主工智能为基础的。我对此额外细目。是以,我认为,你知说念,有好多问题,频频时我听到对于这是否是计较机的一时风俗。你只需要回到基本原则,不雅察本色发生的事情。
东说念主工智能和机器学习的发展相等快,而且影响长远。我在东说念主工智能范围有要紧孝敬的科学家合作的资历,比如多伦多的Alex Krizhevsky、Eliasmith、Hinton和斯坦福的Yan LeCun以及Andrew Ng。、从识别猫咪的肤浅任务到物体识别才智的超东说念主水平的发展,这个经过只用了几年时间。我确信,在畴昔几年内,每个科学范围的每项科学和工程冲突都将以生成性东说念主工智能为基础。饱读动东说念主们不要怀疑这是否只是一时的流行,而应该不雅察本色发生的事情,基于事实来判断。
The computing stack, the way we do computing has changed if the way you write software has changed, I mean, that is pretty cool. Software is how humans encode knowledge. This is how we encode our, you know, our algorithms. We encode it in a very different way. Now that‘s gonna affect everything, nothing else, whatever, be the same. And so I, I think the, the, I think I’m talking to the converted here and we all see the same thing. And all the startups that, you know, you guys work with and the scientists I work with and the engineers I work with, nothing will be left behind. I mean, this, we‘re gonna take everybody with us again.
计较的所有体系,也便是我们进行计较的方式,仍是窜改了,连我们编写软件的方式也窜改了。这意味着我们编码学问的方法也变了,这是一种全新的编码方式。这将会窜改一切,其他的事情都不会和以前一样了。他认为他在这里是对仍是招供这少许的东说念主谈话,群众都看到了相似的趋势。无论是他们合作的初创公司,如故他合作的科学家和工程师,所有东说念主都将被这一变革所影响。他的酷爱是,此次变革将会带领所有东说念主通盘前进。
主合手东说念主:I think one of the most exciting things coming from like the computer science world and looking at all these other fields of science is like I can go to a robotics conference now. Yeah, material science conference. Oh yeah, biotech conference. And like, I‘m like, oh, I understand this, you know, not at every level of the science, but in the driving of discovery, it is all the algorithms that are.
计较机科学范围的一个最令东说念主振作的事情是,当今不错应用于所有其他科学范围。比如,他不错去机器东说念主会议、材料科学会议、生物时刻会议,他会发现我方能妥洽那些内容。固然不是在每个科学范围的每个层面上都懂,但在推动发现方面,都是算法在起作用。
黄仁勋:General and there‘s some universal unifying concepts.
对,有一些普遍澌灭的意见。
主合手东说念主:And I think that‘s like incredibly exciting when you see how effective it is in every domain.
我认为这相等令东说念主振作,当你看到算法在每个范围都如斯有用时。
黄仁勋:Yep, absolutely. And eh, I‘m so excited that I’m using it myself every day. You know, I don‘t know about you guys, but it’s my tutor now. I mean, I, I, I don‘t do, I don’t learn anything without first going to an AI. You know? Why? Learn the hard way. Just go directly to an AI. I should go directly to ChatGPT. Or, you know, sometimes I do perplexity just depending on just the formulation of my questions. And I just start learning from there. And then you can always fork off and go deeper if you like. But holy cow, it‘s just incredible.
我完全快活。我很振作,因为我我方每天都在使用AI。不知你们怎么样,但AI仍是成为我的导师。我当今学任何东西都会先去问AI。为什么?何须要而已去学呢,径直去找AI就行了。比如他会径直去问ChatGPT,或者把柄问题的不同,有时他会去问Perplexity。他会从那里运行学习,然后如果欢乐,不错深入研究。天哪,这真的太不可念念议了。
And almost everything I know, I check, I double check, even though I know it to be a fact, you know, what I consider to be ground truth. I‘m the expert. I’ll still go to AI and check, make double check. Yeah, so great. Almost everything I do, I involve it.
我当今险些作念任何事情都会用到AI。哪怕是他知说念的事实,就算是他是阿谁范围的群众,他也会用AI再搜检一遍。他以为这样很好,因为他险些所有的事情都会让AI参与。
主合手东说念主:I think it‘s a great note to stop on. Yeah, thanks so much that time today.
这是个很好的罢了话题。感谢群众今天的参与,时间到了。
黄仁勋:Really enjoyed it. Nice to see you guys.
我今天很得意见到群众。
风险教唆及免责要求
商场有风险,投资需严慎。本文不组成个东说念主投资建议,也未商量到个别用户特殊的投资方针、财务气象或需要。用户应试虑本文中的任何意见、不雅点或论断是否相宜其特定气象。据此投资,包袱骄贵。
海量资讯、精确解读,尽在新浪财经APP包袱裁剪:陈钰嘉