Good afternoon. My name is David and I'll be your conference operator today. At this time I'd like to welcome everyone to NVIDIA's second quarter earnings call. Today's conference is being recorded. All lines have been placed on me to prevent any background noise. After the speakers are marked, so they are a question and answer session.
If you'd like to ask a question during this time, simply press the star key followed by the number one on your telephone keypad. If you'd like to withdraw your question, press star one once again. Thank you, Samona Jankowski. You may begin your conference.
Thank you. Good afternoon everyone and welcome to NVIDIA's conference call for the second quarter of fiscal 2024. We meet today from NVIDIA, our Jensen Huang, president and chief executive officer and co-ed Cress, executive vice president and chief financial officer.
I'd like to remind you that our call is being webcast live on NVIDIA's investor relations website. The webcast will be available through replay until the conference call to discuss our financial results for the third quarter of fiscal 2024. The content of today's call is NVIDIA's property. It can be reproduced or transcribed without our prior written consent.
During this call, we may make forward-looking statements based on current expectations. These are subject to a number of significant risks and uncertainties and our actual results may differ materially. For discussion of factors that could affect our future financial results and business, please refer to the disclosure in today's earnings release. Our most recent forms, 10K and 10Q, and the reports that we may file on form 8K with the Securities and Exchange Commission. All our statements are made as of today, August 23, 2023, based on information currently available to us. Accept is required by law. We assume no obligation to update any such statements.
During this call, we will discuss non-gap financial measures. We can find a reconciliation of these non-gap financial measures to gap financial measures in our CFO commentary, which is posted on our website. And with that, let me turn the call over to collect.
Thanks, Yamana. We had an exceptional border. Record Q2 revenue of $13.51 billion was up 88% sequentially and up 101% year on year. And above our outlook of $11 billion. Let me first start with data center.
Record revenue of $10.32 billion was up 141% sequentially and up 171% year on year. Data center compute revenue nearly tripled year on year. Driven primarily by accelerating demand for cloud service providers and large consumer internet companies for our HDX platform, the engine of genitive AI and large language models.
Major companies, including AWS, Google Cloud, Meta, Microsoft Azure and Oracle Cloud, as well as growing number of GPU cloud providers are deploying in volume HDX systems based on our Hopper and Ampere architecture Tensor Core GPUs. Networking revenue almost doubled year on year, driven by our end-to-end in Finneban networking platform, the gold standard for AI.
There is tremendous demand for NVIDIA accelerated computing and AI platforms. Our supply partners have been exceptional in ramping capacity to support our needs. Our data center supply chain, including HDX with 35,000 parts and highly complex networking, has been built up over the past decade. We have also developed and qualified additional capacity and suppliers for key steps in the manufacturing process, such as co-host packaging. We expect supply to increase each quarter through next year.
By geography, data center growth was strongest in the US as customers direct their capital investments to AI and accelerated computing. China demand was within the historical range of 20 to 25 percent of our data center revenue, including compute and networking solutions. At this time, let me take a moment to address recent reports on the potential for increased regulations on our exports to China.
We believe the current regulation is achieving the intended results. Given the strength of demand for our products worldwide, we do not anticipate that additional export restrictions on our data center GPUs if adopted would have an immediate material impact toward financial results. However, over the long term, restrictions prohibiting the sale of our data center GPUs to China, if implemented, will result in a permanent loss of an opportunity for the US industry to compete and lead in one of the world's largest markets.
Our cloud service providers drove exceptional strong demand for HDX systems in the quarter as they undertake a generational transition to upgrade their data center infrastructure for the new era of accelerated computing in AI.
The NVIDIA HDX platform is culminating of nearly two decades of full stack innovation across Silicon systems, interconnects, networking, software, and algorithm. Instances powered by the NVIDIA H100 Tensor Core GPUs are now generally available at AWS, Microsoft Azure, and several GPU cloud providers, with others on the way shortening.
Consumer Internet companies also drove the very strong demand. Their investments in data center infrastructure, purpose built for AI, are already generating significant returns. For example, meta recently highlighted that since launching reels and AI recommendations have driven a more than 24% increase in time spent on Instagram.
Enterprises are also racing to deploy generative AI, driving strong consumption of NVIDIA powered instances in the cloud as well as demand for on-premise infrastructure. Whether we serve customers in the cloud or on-prem, through partners or DRAP, their applications can run seamlessly on NVIDIA AI enterprise software with access to our acceleration libraries, retain models, and APIs.
We announced a partnership with Snowflake to provide enterprises with accelerated paths to create customized generative AI applications using their own proprietary data, all securely within the Snowflake data cloud. With the NVIDIA Nemo platform for developing large language models, enterprises will be able to make custom LLMs for advanced AI services, including chat box, search, and summarization right from the Snowflake data cloud.
Virtually every industry can benefit from generative AI. For example, AI co-pilots, such as those just announced by Microsoft, and boost the productivity of over a billion office workers and tens of millions of software engineers. Millions of professionals in legal services, sales, customer support, and education will be available to leverage AI systems trained in their fields. AI co-pilots and assistants are set to create new multi-hundred billion dollar market opportunities for our customers.
We are seeing some of the earliest applications of generative AI in marketing, media, and entertainment. WPP, the world's largest marketing and communication services organization, is developing a content engine using NVIDIA Amiverse to enable artists and designers to integrate generative AI into 3D content creation. WPP designers can create images from text prompts while responsibly trained generative AI tools and content from NVIDIA partners, such as Adobe and Getty Images, using NVIDIA Picasso, a foundry for custom generative AI models for visual design. All content provider, Shutter Stott, is also using NVIDIA Picasso to build tools and services that enable users to create 3D scene backgrounds with the help of generative AI.
We partnered with ServiceNow, an Accenture, to launch the AI Lighthouse program, fast tracking the development of enterprise AI capabilities. AI Lighthouse Unites with ServiceNow Enterprise Automation Platform and Engine with NVIDIA Accelerated Computing and with Accenture Consulting and Deployment Services. We are collaborating also with HuggingFace to simplify the creation of new and custom AI models for enterprises. HuggingFace will offer a new service for enterprises to train and to advanced AI models powered by NVIDIA DGX Cloud.
And just yesterday, VMware and NVIDIA announced a major new enterprise offering called VMware Private AI Foundation with NVIDIA, a fully integrated platform featuring AI software and Accelerated Computing from NVIDIA with multi-cloud software for enterprises running VMware. VMware is hundreds of thousands of enterprise customers who have access to the infrastructure. AI and Cloud Management software needed to customize models and run generative AI applications such as intelligent chat box, assistance, search and some organizations.
就在昨天,VMware和NVIDIA宣布了一项名为VMware Private AI Foundation with NVIDIA的重要企业服务,这是一个完全集成的平台,结合了来自NVIDIA的AI软件和加速计算,并提供了适用于运行VMware的企业的多云软件。VMware拥有数十万的企业客户,他们可以使用基础设施、AI和云管理软件来定制模型并运行生成式AI应用,如智能聊天框、辅助工具、搜索等。
We also announced new NVIDIA AI Enterprise Ready Servers featuring the new NVIDIA L40S GPU built for the industry standard data center server ecosystem and Bluefield 3 DPU data center infrastructure processor. L40S is not limited by co-law supply and is shipping to the world leading server system makers. L40S is a universal data center processor designed for high volume data center scanning out to accelerate the most compute intensive applications including AI training and infancy, 3D design and visualization, video processing and NVIDIA Amigurz industrial digitalization.
NVIDIA AI Enterprise Ready Servers are fully optimized for VMware, cloud foundation and private AI foundation. Nearly 100 configurations of NVIDIA AI Enterprise Ready Servers will soon be available from the world leading enterprise IT computing companies including Dell, HPE and Lenovo.
NVIDIA AI Enterprise Ready Servers已经完全优化了VMware、云基础设施和私有AI基础设施。世界领先的企业IT计算公司,包括戴尔、惠普和联想,很快将提供近100种配置的NVIDIA AI Enterprise Ready Servers。
The GH 200 Grace Hopper Superchip, which combines our ARM based Grace CPU with Hopper GPU, entered full production and will be available this quarter in OEM servers. It is also shipping to multiple super computing customers including Los Altmos, National Lab, and the Swiss National Computing Center. NVIDIA and SoftBank are collaborating on a platform based on GH 200 for generative AI and 5G6G applications.
The second generation version of our Grace Hopper Superchip with the latest HBM3E memory will be available in Q2 of calendar 20-24. We announced the GGX GH 200, a new class of large memory AI supercomputer for giant AI language model, recommendations for systems, and data analytics. This is the first use of the new NVIDIA NV-based Twitch system enabling all of its 256 Grace Hopper Superchip to work together as one, a huge jump in territorial fire generation connecting just HEPUs over NV-based.
The GH 200 systems are expected to be available at the end of the year. Google Cloud, Meta, and Microsoft are among the first to gain access. Strong networking growth was driven primarily by the Finneban infrastructure to connect HGX GPU systems, thanks to its end-to-end optimization and in-network computing capabilities. Finneban delivers more than double the performance of traditional Ethernet for AI. For billions of dollar AI infrastructures, the value from the increased throughput of the Finneban is worth hundreds of millimals and paid for the network. In addition, only in Finneban can scale to hundreds of thousands of GPUs. It is the network of choice for leading AI practitioners.
For Ethernet-based cloud data centers that seek to optimize their AI performance, we announced NVIDIA Spectrum X, an accelerated networking platform designed to optimize Ethernet for AI workflows. Spectrum X couples the spectrum for Ethernet switch with the Bluefield 3 DPU achieving 1.5x better overall AI performance and power efficiency versus traditional Ethernet. Bluefield 3 DPU is a major success. It is in qualification with major OEMs and ramping across multiple CSPs and consumer network companies.
Now moving to gaming. Gaming revenue of 2.49 billion was up 11% sequentially and 22% year-on-year. Growth was fueled by GeForce RTX 40 Series GPUs from laptops and desktop. Any customer demand was solid and consistent with seasonality. We believe global end demand was returned to growth after last year's slowdown. We have a large upgrade opportunity ahead of us. Just 47% of our installed base have upgraded to RTX and about 20% of a GPU with an RTX 3060 or higher performance.
Laptop GPUs posted strong growth in the key back-to-school season led by RTX 4060 GPUs. NVIDIA's GPU powered laptops have gained in popularity and their shipments are now outpacing desktop GPUs from several regions around the world. This is likely to shift the reality of our overall gaming revenue events. We have Q2 and Q3 as the stronger quarters of the year, reflecting the back-to-school and holiday build schedules for laptops. In desktop, we launched the GeForce RTX 4060 and the GeForce RTX 4060 TIGPAs, bringing the 8-in-lub-lace architecture down to price points as low as $299. The ecosystem of RTX and DLSS games continue to expand. 35 new games added to DLSS support, including blockbusters such as Diablo 4 and Baldur's Gate 3. There's now over 330 RTX accelerated games and apps. We are bringing generative AI to games, a Computex-reannounced NVIDIA Avatar Cloud Engine for games a custom AI model-friendly service. Developers can use this to bring intelligence to non-player characters. It enhances a number of NVIDIA on-nubers and AI technology, including Nemo, Reba and Audio 2-Face.
Now moving to professional visualization. Revenue of $375 million was up 28% sequentially and down 24% year-on-year. The ADA architecture ramp drove strong growth into two, rolling out initially the laptop workstations with a refresh of desktop workstations coming into three. These will include powerful new RTX systems with up to four NVIDIA RTX 6000 GPUs, providing more than 5,800 teraflops of AI performance and 192 gigabytes of GPU memory. They can be configured with NVIDIA AI Enterprise or NVIDIA Baldur's Enterprise. We also announced three new desktop workstations with GPUs based on the ADA generation. The NVIDIA RTX 5000, 4500, and 4000, offering up to 2X the RT Core throughput and up to 2X faster AR training performance compared to the previous generation.
现在转向专业可视化。第四季度收入为3.75亿美元,环比增长28%,同比下降24%。ADA架构的推出推动了业务的强劲增长,首先推出了笔记本工作站,并在第三季度推出更新的台式工作站。这些机型包括提供超过5800万AI计算性能和192GB GPU内存的强大新型RTX系统,可配置为NVIDIA AI Enterprise或NVIDIA Baldur's Enterprise。我们还宣布推出基于ADA一代的三款台式工作站,搭载了NVIDIA RTX 5000、4500和4000 GPU,相比上一代,其新一代RT核心吞吐量提高了2倍,AR训练性能提高了2倍。
In addition to the traditional workloads such as 3D Design and Content Creation, new workloads in generative AI, large language model development, and data science are expanding the opportunity in pro visualization for our RTX technology.
One of the key themes in Jensen's keynote at SIGGRAPH earlier this month was the conversion of graphics and AI. This is where NVIDIA on-nubers is positioned. On-nubers is OpenUSD's native platform. OpenUSD is a universal interchange that is quickly becoming the standard for the 3D world, much like HTML is the universal language for the 2D and it's best. Together Adobe, Apple, Autodesk, Pixar and NVIDIA forms the Alliance for OpenUSD. Our mission is to accelerate OpenUSD's development and adoption.
Moving to automotive. Revenue was 253 million, down 15% sequentially and up 15% year on year. Solid year on year growth was driven by the ramp of self-driving platforms based on the renewable drive or an SOC with a number of new energy vehicle makers. These sequential declines reflect lower overall automotive demand, particularly in China.
We announced a partnership with MediaTac to bring drivers and passenger new experiences inside the call. MediaTac will develop automotive SOCs and integrate a new product line of NVIDIA GPU chiplets. The partnership covers a wide range of vehicle segments from luxury to entry novel.
Moving to the rest of the P&L, gap gross margins expanded to 70.1% and non-gap gross margins to 71.2% driven by higher data center sales. Our data center products include a significant amount of software and complexity which is also helping thrive our gross margins. Sequential gap operating expenses were up 6% and non-gap operating expenses were up 5% primarily reflecting increased compensation and benefits.
We return approximately 3.4 billion to shareholders in the form of share repurchases and cash dividends. Our board of directors has just approved an additional 25 billion in stock repurchases to add to our remaining 4 billion of authorization as of the end of Q2.
That may turn to the outlook for the third quarter of fiscal 2024. Demand for our data center platform for AI is tremendous and broad based across industries and customers. Our demand visibility extends into next year. Our supply over the next several quarters will continue to ramp as we lower cycle times and work with our supply partners to add capacity. Additionally, the new L40S GPU will help address the growing demand for many types of workloads from cloud to enterprise.
For Q3, total revenue is expected to be 16 billion plus or minus 2%. We expect sequential growth to be driven largely by data center with gaining and provision also contributing. Gap and non-gap gross margins are expected to be 71.5% and 72.5%, respectively, plus or minus 50 basis points. Gap and non-gap operating expenses are expected to be approximately 2.95 billion and 2 billion respectively. Gap and non-gap other income and expenses are expected to be an income of approximately 100 billion, excluding gains and losses from non-affirming investments. Gap and non-gap tax races are expected to be 14.5% plus or minus 1%, excluding any discrete items.
Further financial details are included in the CFO commentary and other information available on our IR website.
进一步的财务细节包括在首席财务官的评论和我们投资者关系网站上提供的其他信息中。
In closing, let me highlight some upcoming events for the financial community. We will attend the Jeffrey's Tech Summit on August 30th in Chicago, the Goldman Sachs Tech Conference on September 5th in San Francisco, the Evercore Summit Conduct conference on September 6th, as well as the City Tech Conference on September 7th, both in New York and the B&A Virtual AI Conference on September 7th.
最后,让我强调一下金融界即将举行的一些重要活动。我们将参加8月30日在芝加哥举行的杰弗里斯科技峰会,9月5日在旧金山举行的高盛技术会议,9月6日的 Evercore 峰会行为会议,以及9月7日在纽约举行的 City Tech 会议,以及9月7日的 B&A 虚拟人工智能会议。
Our earnings call to discuss the results of our third quarter of the school 2024 is scheduled for Tuesday, November 1st.
我们将于2024年第三学期的收益电话会议定于11月1日星期二,讨论我们的业绩结果。
Operator, we will now open the call for questions. Could you please call for questions for us? Thank you.
操作员,我们现在开始接受问题了。你能否代为我们提问?谢谢。
At this time, I would like to remind everyone in order to ask a question, press star than the number one on your telephone keypad. We ask that you please limit yourself to one question.
Yes, thank you very much. Good afternoon. Obviously, remarkable results. Jensen, I wanted to ask a question of you regarding the really quickly emerging application of large model inference. I think it is pretty well understood by the majority of investors that you guys have a very much a lockdown share of the training market. A lot of the smaller model inference workloads have been done on A6 or CPUs in the past. With many of these GPT and other really large models, there is this new workload that is accelerating super duper quickly on large model inference. I think your Grace Hopper super chip products and others are pretty well aligned for that. Could you maybe talk to us about how you are seeing the inference market segment between small model inference and large model inference and how your product portfolio is positioned for that? Thanks.
Yeah, thanks a lot. Let's take a quick step back. These large language models are fairly pretty phenomenal. It does several things, of course. It has the ability to understand unstructured language. At its core, what it has learned is the structure of human language and it has encoded within it compressed within it a large amount of human knowledge that has learned by the corpus that it has studied. What happens is you create these large language models and you create as large as you can and then you derive from it smaller versions of the model. Essentially, teacher student models. It is a process called distillation.
When you see these smaller models, it is very likely the case that they were derived from or distilled from or learned from larger models. Just as you have professors and teachers and students and so on and so forth. You are going to see this going forward. You start from a very large model and it has a large amount of generality and generalization and what is called zero shot capability. For a lot of applications and questions or skills that you haven't trained it specifically on, these large language models miraculously has the capability to perform them. That is what makes it so magical. On the other hand, you would like to have these capabilities in all kinds of computing devices. What you do is you distill them down. These smaller models might have excellent capabilities on a particular skill but they don't generalize as well. They don't have what is called as good zero shot capabilities. They all have their own unique capabilities but you start from very large models.
Next we will go to Vivek Aria with B of A security. Your line is now open. Thank you. Just had a quick clarification on our question. Collet, if you could please clarify how much incremental supply do you expect to come online in the next year? You think it is up 20, 30, 40, 50 percent? So just any sense of how much supply because you said it is growing every quarter. Then, Jensen, the question for you is when we look at the overall hyperscaler spending, that pie is not really growing that much. What is giving you the confidence that they can continue to carve out more off that pie for a generative AI? Give us your sense of how sustainable is this demand as we look over the next one to two years. If I take your implied Q3 outlook of data center, 12, 13 billion, what does that say about how many servers are already AI accelerated? Where is that going? Give us some confidence that the growth that you are seeing is sustainable into the next one to two years.
接下来,我们将与B of A证券一起前往Vivek Aria。你现在可以发言了。谢谢您。我们对我们的问题有一个快速澄清。Collet,请问你认为明年会有多少额外的供应量上线?你认为它会增长20、30、40、50%吗?所以只要有个大致的供应量,因为你说它每个季度都在增长。然后,Jensen,你的问题是当我们看整体的超大规模支出时,那个饼图并没有真正增长那么多。是什么让你有信心他们能继续从那个饼图中分得更多用于生成AI的份额?请告诉我们您对未来一到两年的需求的可持续性的看法。如果我们根据你所暗示的Q3数据中心展望进行估算,12,13亿美元,这意味着已经有多少服务器已经进行了AI加速?这会导致什么结果?请提供一些使我们相信您所看到的增长在未来一到两年是可持续的。
Thanks for that question regarding our supply. Yes, we do expect to continue increasing, ramping our supply over the next quarter as well as into next fiscal year. In terms of percent, that is not something that we have here. It is a work across so many different suppliers, so many different parts of building and HDX and many of our other new products that are coming to market. But we are very pleased with both the support that we have with our suppliers and the long time that we have spent with them in improving their supply.
The world has something along the lines of about a trillion dollars worth of data centers installed in the cloud and enterprise and otherwise. That trillion dollars of data centers is in the process of transitioning into accelerated computing and generative AI.
We are seeing two simultaneous platform shifts at the same time. One is accelerated computing and the reason for that is because it is the most cost effective, most energy effective and the most performant way of doing computing now.
What you are seeing, and then all of a sudden, enabled by generative AI, enabled by accelerated computing, generative AI came along. This incredible application now gives everyone two reasons to transition to do a platform shift from general purpose computing, the classical way of doing computing to this new way of doing computing, accelerated computing.
It is about a trillion dollars worth of data centers. Call it a quarter of a trillion dollars of capital spend each year. You are seeing that data centers around the world are taking that capital spend and focusing it on the two most important trends of computing today, accelerated computing and generative AI. I think this is not a near term thing. This is a long term industry transition and we are seeing these two platforms shifts happening at the same time.
Next we go to Stacy Raskon with Bernstein research. Hi guys. Thanks for taking my question. I was wondering, Collette, if you could tell me how much of data center in the quarter, maybe even the guidance like systems versus GPT, like DGX versus just the H100. What I am really trying to get at is how much is pricing or content or everyone to define that versus unit driving the growth going forward. Can you give us any color around that?
Sure Stacy. Let me help within the quarter. Our HDX systems are a very significant part of our data center as well as our data center growth that we had seen. Those systems include our HDX of our helper architecture but also our amper architecture. Yes, we are still selling both of these architectures or in the market.
When you think about that, what does that mean from both the systems as a unit of course is growing quite substantially and that is driving in terms of the revenue increases. All of these things are the drivers of the revenue inside data center. Our DGXs are always a portion of additional systems that we will sell. Those are great opportunities for enterprise customers and many other different types of customers that we are seeing even in our consumer internet companies.
The importance of there is also coming together with software that we sell with our DGXs but that is a portion of our sales that we are doing. The rest of the GPUs, we have new GPUs coming to market that we talk about the L40S and they will add continued growth going forward but again the largest driver of our revenue within this last quarter was definitely the HDX system.
And Stacy, if I could just add something, you say it is H100 and I know you know what your mental image in your mind but the H100 is 35,000 parts, 70 pounds, nearly a trillion transistors in combination. It takes a robot to build, well many robots to build because it is 70 pounds to lift and it takes a supercomputer to test a supercomputer. So these things are technology marvels and the manufacturing of them is really intensive. So I think we call it H100 as if it is a chip that comes off of a fab but H100 goes out really as HGX and it is in the world's hyperscalers and they are really quite large system components if you will.
Next we go to Mark Lipikis with Jeffries. Your line is no open. Hi, thanks for taking my question and congrats on the success. Jensyn, it seems like a key part of the success, your success in the market is delivering the software ecosystem along with the chip and the hardware platform and I had a two-part question on this. I was wondering if you could just help us understand the evolution of your software ecosystem, the critical elements and is there a way to quantify your lead on this dimension, like how many person years you have invested in building it and then part two, I was wondering if you would care to share with us your view on what percentage of the value of the NVIDIA platform is hardware differentiation versus software differentiation. Thank you.
Mark, I would really appreciate the question. Let me see if I could use some metrics. So we have a run time called NVIDIA AI Enterprise. This is one part of our software stack and this is, if you will, the run time that just about every company uses for the end-to-end of machine learning, from data processing, the training of any model that you like to do on any framework you like to do, the inference and the deployment, the scaling it out into a data center. It could be a scale out for a hyperscale data center, it could be a scale out for enterprise data center, for example, on VMware. You can do this on any of our GPUs. We have hundreds of millions of GPUs in the field and millions of GPUs in the cloud and just about every single cloud. It runs in a single GPU configuration as well as multi-GPU per compute or multi-node. It also has multiple sessions or multiple computing instances per GPU. So from multiple instances per GPU to multiple GPUs, multiple nodes to entire data center scale.
马克,我真的很感激这个问题。让我看看是否能使用一些指标来回答。所以我们有一个叫做NVIDIA AI Enterprise的运行时。这是我们软件堆栈的一部分,它可以说是每家公司在机器学习的端到端过程中都会使用的运行时,包括数据处理、训练任何你喜欢使用的模型在任何框架上、推理和部署,以及将其扩展到数据中心。它可以用于超大规模数据中心的扩展,也可以用于企业数据中心(例如在VMware上)。你可以在我们的任何GPU上进行这些操作。我们在现场有数亿个GPU,云上有数百万个GPU,几乎在每个云平台都有。它可以在单个GPU配置下运行,也可以在多个GPU或多节点上进行计算,甚至可以实现每个GPU的多个会话或多个计算实例。所以从每个GPU的多个实例到多个GPU,多个节点再到整个数据中心的规模,它都可以胜任。
So this run time called NVIDIA AI Enterprise has something like 4500 software packages, software libraries and has something like 10,000 dependencies among each other. That run time is, as I mentioned, continuously updated and optimized for our install base, for our stack. That's just one example of what it would take to get accelerated computing to work. The number of code combinations and type of application combinations is really quite insane. That's taken us two decades to get here.
所以这个名为NVIDIA AI Enterprise的运行时有大约4500个软件包、软件库,并且它们之间有大约10000个依赖关系。正如我所提到的,这个运行时在我们的安装基础和堆栈上不断进行更新和优化。这只是让加速计算起作用所需的一个例子。代码组合和应用组合的数量真的相当可怕。我们花了20年的时间才到达这一点。
But what I would characterize as probably the elements of our company, if you will, are several. I would say number one is architecture. The flexibility, the versatility and the performance of our architecture makes it possible for us to do all the things that I just said. From data processing to training to inference, pre-processing of the data before you do the inference to the post-processing of the data, tokenizing of languages so that you could then train with it. The workflow is much more intense than just training or inference. But anyways, that's where focus and it's fine.
But when people actually use these computing systems, it requires a lot of applications. So the combination of our architecture makes it possible for us to deliver the lowest cost of ownership. The reason for that is because we accelerate so many different things.
The second characteristic of our company is the install base. You have to ask yourself, why is it that all the software developers come to our platform? The reason for that is because software developers seek a large install base so that they can reach the largest number of end users so that they could build the business or get a return on the investments that they make.
And then the third characteristic is reach. We're in the cloud today, both for public cloud, public-facing cloud because we have so many customers that use it, so many developers and customers that use our platform. CSPs are delighted to put it up in the cloud. They use it for internal consumption to develop and train and to operate recommender systems or search or data processing engines and whatnot all the way to training and inference. So we're in the cloud, we're in enterprise. Yesterday we had a very big announcement. It's really worthwhile to take a look at that. VMware is the operating system of the world's enterprise. We've been working together for several years now and we're going to bring together, together we're going to bring generative AI to the world's enterprises all the way out to the edge. And so reach is another reason. And because of reach, all of the world's system makers are anxious to put NVIDIA's platform in their systems. And so we have a very broad distribution from all of the world's OEMs and ODMs and so on and so forth because of our reach.
And then lastly because of our scale and velocity. We were able to sustain this really complex stack of software and hardware networking and compute and across all of these different usage models and different computing environments. And we're able to do all this while accelerating the velocity of our engineering. It seems like we're introducing a new architecture every two years. Now we're introducing a new architecture, a new product just about every six months. And so these properties make it possible for the ecosystem to build their company and their business on top of us. And so those in combination makes it special.
Next we'll go to Atif Malik with Citi. Your line's open. Hi, thank you for taking my question and great job on results in Atkuk. I have a question on the Kowals less L for TS that you guys talked about. Any idea how much of the supply tightness can L40 as help with. And if you can talk about the incremental profitability or growth margin contribution from this product. Thank you.
接下来我们将与花旗银行的Atif Malik进行对话。请问你有什么问题?谢谢你回答我的问题,对Atkuk的成果表现也很出色。我想问一下关于你们提到的Kowals less L for TS,你们有多少转供应紧张情况可以由L40来帮助解决呢?同时,你能谈一谈这个产品对增量利润或者增长利润率的贡献吗?谢谢。
Yeah, Atif let me take that for you. The L40, L40s is really designed for a different type of application. H100 is designed for large scale language models and processing just very large models and a great deal of data. And so that's not L40s' focus. L40s' focus is to be able to fine tune models, fine tune pre-trained models. And it will do that incredibly well. It has a transformer engine, it's got a lot of performance. You can get multiple GPUs in a server. It's designed for hyperscale scale out. Meaning it's easy to install L40s servers into the world's hyperscale data centers. It comes in a standard rack, standard server, and everything about it is standard. And so it's easy to install. L40s also is with the software stack around it and along with Bluefield 3. And although the work that we did with there, and the work that we did with Snowflakes and ServiceNow and so many other enterprise partners, L40s is designed for the world's enterprise IT systems. And that's the reason why HPE, Dell, and Lenovo and some other 20 other system makers building about 100 different configurations of enterprise servers are going to work with us to take generative AI to the world's enterprise. And so L40s is really designed for a different type of scale out, if you will. It's of course, large language models. It's of course generative AI. But it's a different use case. And so the L40s is off to a great start. And the world's enterprise and hyperscalers are really clamoring to get L40s deployed.
Okay, next we'll go to Joe Moore with Morgan Stanley. The line is now open. Great. Thank you. I guess the thing about these numbers that's so remarkable to me is the amount of demand that remains unfulfilled, talking to some of your customers. As good as these numbers are, you sort of more than tripled your revenue in a couple of quarters. There's a demand in some cases for multiples of what people are getting. So can you talk about that, how much unfulfilled demand do you think there is? And you talked about visibility extending into next year, you know, if you have line of sight into when you'll get to supply demand equilibrium here.
Yeah, we have excellent visibility through the year and into next year. And we're already planning the next generation infrastructure with the leading CSPs and data center builders. The demand, the easiest way to think about the demand is the world is transitioning from general purpose computing to accelerated computing. That's the easiest way to think about the demand. The best way for companies to increase their throughput, improve their energy efficiency, improve their cost efficiency is to divert their capital budget to accelerated computing and generous to AI. Because by doing that, you're going to offload so much work to upload off of the CPUs. That the available CPUs is in your data center will get boosted. And so what you're seeing companies do now is recognizing this tipping point here, recognizing the beginning of this transition and diverting their capital investment to accelerated computing and generous to AI. And so that's probably the easiest way to think about the opportunity ahead of us. This isn't a singular application that is driving the demand, but this is a new computing platform, if you will, a new computing transition that's happening. And data centers all over the world are responding to this and shifting in a broad-based way.
Next we go to Shia Hari with Goldman Sachs. Your line is now open. Hi, thank you for taking the question. I had one quick clarification question for Collette and then another one for Jensen. Collette, I think last quarter you had said CSPs were about 40% of your data center revenue. Consumer Internet 30% enterprise 30%. Based on your remarks, it sounded like CSPs and consumer internet may have been a larger percentage of your business. If you can kind of clarify that or confirm that, that would be super helpful. And then Jensen, a question for you, given your position as the key enabler of AI, the breadth of engagements and the visibility you have into customer projects, I'm curious how confident you are that there will be enough applications or use cases for your customers to generate reasonable return on their investments. I guess I asked the question because there is a concern out there that there could be a bit of a pause in your demand profile in the out years. Curious if there is enough breadth and depth there to support a sustained increase in your data center business going forward. Thank you.
Okay, so thanks, Toshiya, on the question regarding our types of customers that we have in our data center business. And we look at it in terms of combining our compute as well as our networking together. Our CSPs, our large CSPs are contributing a little bit more than 50% of our revenue within Q2. And the next largest category will be our consumer internet companies. And then the last piece of it will be our enterprise and a high performance computing.
Toshiya, I'm reluctant to guess about the future. And so I'll answer the question from the first principle of computer science perspective. It is recognized for some time now that general-purpose computing is just not in brute forcing general-purpose computing. Using general-purpose computing at scale is no longer the best way to go forward. It's too energy costly. It's too expensive. And the performance of the applications is too slow. And finally, the world has a new way of doing it. It's called accelerated computing. And what kicked it into turbo charge is generous of AI. But accelerated computing could be used for all kinds of different applications that are already in the data center. And by using it, you offload the CPUs. You save a ton of money, you use an order of magnitude in cost and order of magnitude in energy and the throughput is higher. And that's what the industry is really responding to. Going forward, the best way to invest in the data center is to divert the capital investment from general-purpose computing and focus it on generative AI and accelerated computing.
Generative AI provides a new way of generating productivity, a new way of generating new services to offer to your customers. And accelerated computing helps you save money and save power. And the number of applications is tons. Lots of developers, lots of applications, lots of libraries. It's ready to be deployed. And so I think the data centers around the world recognize this. This is the best way to deploy resource and deploy capital going forward for data centers. And this is true for the world's clouds and you're seeing a whole crop of new GPU specialty, GPU specialized cloud service providers. One of the famous ones is CoreWeave and they're doing incredibly well. But you're seeing regional GPU specialists, service providers all over the world now. And it's because they all recognize the same thing, that the best way to invest your capital going forward is to put it to already computing and generative AI.
We're also seeing that enterprises want to do that. But in order for enterprises to do it, you have to support the management system, the operating system, the security and software-defined data center approach of enterprises and that's called VMware. And we've been working several years with VMware to make it possible for VMware to support not just the virtualization of CPUs but the virtualization of GPUs as well as the distributed computing capabilities of GPUs supporting NVIDIA's Bluefield for high-performance networking. And all of the generative AI libraries that we've been working on is now going to be offered as a special skew by VMware's Salesforce, which is, as we all know, quite large because they reach some several hundred thousand VMware customers around the world. And this new skew is going to be called VMware Private AI Foundation. And this will be a new skew that makes it possible for enterprises. And in combination with HP Dell and Lenovo's new server offerings based on L40S, any enterprise could have a state-of-the-art AI data center and be able to engage and enter to AI. And so I think the answer to that question is hard to predict exactly what's going to happen quarter to quarter. But I think the trend is very, very clear now that we're seeing a platform shift.
我们也看到企业希望这样做。但是为了让企业能够做到这一点,您必须支持企业的管理系统、操作系统、安全性和软件定义的数据中心方法,这就是所谓的VMware。我们已经与VMware合作了几年,使其能够不仅支持CPU的虚拟化,还支持GPU的虚拟化以及支持NVIDIA的Bluefield用于高性能网络的分布式计算能力。我们一直在研发的所有AI生成库现在将作为VMware Salesforce的特殊外设提供,众所周知,VMware Salesforce非常庞大,因为它们拥有全球数十万VMware客户。这个新的外设将被称为VMware Private AI Foundation。这将是一个使企业能够实现的新的外设。结合基于L40S的惠普戴尔和联想的新服务器提供,任何企业都可以拥有先进的AI数据中心,并能够参与和进入AI。因此,我认为对这个问题的答案很难准确预测每个季度会发生什么。但我认为现在已经非常清楚,我们正在见证一个平台转变的趋势。
Next we'll go to Timothy R. Curie with UBS. Your line is now open. Thanks a lot. Can you talk about the attach rate of your networking solutions to the compute that you're shipping? In other words, is like half of your compute shipping with your networking solutions, you know, more than half, less than half? And is this something that maybe you can use to prioritize allocation of the GPUs? Thank you.
Well, working backwards, we don't use that to prioritize the allocation of our GPUs. We let customers decide what networking they would like to use. And for the customers that are building very large infrastructure, Infiniband is, you know, I hate to say it, kind of a no-brainer. And the reason for that, because the efficiency of Infiniband is so significant, you know, some 10, 15, 20 percent higher throughput for a billion dollars infrastructure translates to enormous savings. Basically, the networking is free. And so if you have a single application, if you will, infrastructure, or it's largely dedicated to large language models or large AI systems, Infiniband is really a terrific choice.
However, if you're hosting for a lot of different users and Ethernet is really according to the way you manage your data center, we have an excellent solution there that we just recently announced and it's called Spectrum X. Well, we're going to bring the capabilities, if you will, not all of it, but some of it of the capabilities of Infiniband to Ethernet. And so that we can also, within the environment of Ethernet, allow you to enable you to get excellent generative AI capabilities. So Spectrum X is just ramping now. It requires Bluefield 3 and it supports both our Spectrum 2 and Spectrum 3 Ethernet switches and the additional performance is really spectacular. And Bluefield 3 makes it possible and a whole bunch of software that goes along with it. Bluefield, as all of you know, is a project really dear to my heart and it's off to just a tremendous start. I think it's a home run. And this is the concept of in-network computing and putting a lot of software in the computing fabric is being realized with Bluefield 3 and it is going to be a home run.
Our final question comes from the line of Ben Ritzis with Millius. Your line is no open. Hi, good afternoon, good evening. Thank you for the question from putting me in here. My question is with regard to DGX Cloud. Can you talk about the reception that you're seeing and how the momentum is going and then collect? Can you also talk about your software business? What is the run rate right now and the materiality of that business? And it does seem like it's already helping margins a bit. Thank you very much.
DGX Cloud's strategy. Let me start there. DGX Cloud's strategy is to achieve several things. Number one, to enable a really close partnership between us and the world CSPs. We recognize that many of our, well, we work with some 30,000 companies around the world. 15,000 of them are startups. Thousands of them are generative AI companies. The fastest growing segment, of course, is generative AI. We're working with all of the world's AI startups. And ultimately, they would like to be able to land in one of the world's leading clouds. And so we built DGX Cloud as a footprint inside the world's leading clouds so that we could simultaneously work with all of our AI partners and help land them in easily in one of our cloud partners.
The second benefit is that it allows our CSPs and ourselves to work really closely together to improve the performance of hyperscale clouds, which is historically designed for multi-tenancy and not designed for high-performance distributed computing live generative AI.
And so to be able to work closely architecturally to have our engineers work hand-in-hand to improve the networking performance and the computing performance has been really powerful, really terrific.
因此,能够在构架上与工程师紧密合作,让他们携手改善网络性能和计算性能,真的非常有力量,非常棒。
And then thirdly, of course, NVIDIA uses very large infrastructures ourselves. And our self-driving car team, our NVIDIA research team, our generative AI team, our language model team, the amount of infrastructure that we need is quite significant.
None of our optimizing compilers are possible without our DGX systems. Even compilers these days require AI and optimizing software and infrastructure software requires AI to even develop.
It's been well publicized that our engineering uses AI to design our chips. And so the internal, our own consumption of AI, our robotics team, so on and so forth, our omnivores team, so on and so forth, all needs AI. So our internal consumption is quite large as well and we land that in DGX cloud.
And so DGX cloud has multiple use cases, multiple drivers. And it's been off to just an enormous success and our CSPs love it. The developers love it. And our own internal engineers are clamoring to have more of it. And it's a great way for us to engage and work closely with all of the AI ecosystem around the world.
And let's see if I can answer your question regarding our software revenue. In part of our opening remarks that we made as well, the number of software is a part of almost all of our products, whether they are data center, products, GPU system, or any of our products within gaming and our future automotive products.
You're correct. We're also selling it in extendable business. And that stand alone software continues to grow where we are providing both the software services upgrades across there as well.
Now we're seeing at this point probably hundreds of millions of dollars annually for software business. And we are looking at NVIDIA AI enterprise to be included with many of the products that we're selling, such as our DGX, such as our PCIe versions of our H100.
And I think we're going to see more availability with our CSP marketplaces. So we're up to a great start and I do believe we'll see this continue to grow going forward.
A new computing era has begun. The industry is simultaneously going through two platform transitions, accelerated computing and generative AI.
一个新的计算时代已经开始。该行业正在同时经历两个平台转型,加速计算和生成式人工智能。
Data centers are making a platform shift from general purpose to accelerated computing. The trillion dollars of global data centers will transition to accelerated computing to achieve an order of magnitude better performance, energy efficiency and cost.
Accelerated computing enabled generative AI, which is now driving a platform shift in software and enabling new, never before possible applications. Together, accelerated computing and generative AI are driving a broad-based computer industry platform shift.
Our demand is tremendous. We are significantly expanding our production capacity. Supply will substantially increase for the rest of this year and next year. NVIDIA has been preparing for this for over two decades and has created a new computing platform that the world's industry, world's industries can build upon.
What makes NVIDIA special are one, architecture. NVIDIA accelerates everything from data processing, training, inference, every AI model, real-time speech to computer vision, and giant recommenders to vector databases. The performance and versatility of architecture translates to the lowest data center TCO and best energy efficiency.
Two, install base. NVIDIA has hundreds of millions of CUDA compatible GPUs worldwide. Developers need a large install base to reach end users and grow their business. NVIDIA is the developer's preferred platform. More developers create more applications that make NVIDIA more valuable for customers.
Three, reach. NVIDIA is in clouds, enterprise data centers, industrial edge, PCs, workstations, instruments, and robotics. Each has fundamentally unique computing models and ecosystems. Systems suppliers, like OEMs, computer OEMs, can confidently invest in NVIDIA because we offer significant market demand and reach.
Scale and velocity. NVIDIA has achieved significant scale and is 100% invested in accelerated computing and generative AI. Our ecosystem partners can trust that we have the expertise, focus, and scale to deliver a strong roadmap and reach to help them grow.
We are accelerating because of the additive results of these capabilities. We are upgrading and adding new products about every six months versus every two years to address the expanding universe of generative AI. While we increase the output of H100 for training and inference of large language models, we are ramping up our new L40S universal GPU for cloud scale out and enterprise servers.
Spectrum X, which consists of our Ethernet switch, Bluefield 3 SuperNIC, and software helps customers who want the best possible AI performance on Ethernet infrastructures. Customers are already working on next generation accelerated computing and generative AI with our Grace Hopper.
We are extending NVIDIA AI to the world's enterprises, that demand generative AI, but with the model privacy, security, and sovereignty. Together with the world's leading enterprise IT companies, Accenture, Adobe, Getty, Hugging Face, Snowflake, ServiceNow, VMware, and WPP, and our enterprise system partners Dell, HPE, and Lenovo, we are bringing generative AI to the world's enterprise.
We're building NVIDIA Omniverse to digitalize and enable the world's multi-trillion dollar heavy industries to use generative AI to automate how they build and operate physical assets and achieve greater productivity. Generative AI starts in the cloud, but the most significant opportunities are in the world's largest industries, where companies can realize trillions of dollars of productivity gains.
It is an exciting time for NVIDIA, our customers, partners, and the entire ecosystem to drive this generational shift in computing. We look forward to updating you on our progress next quarter. This includes today's conference call. You may now disconnect.