This conference call transcript was computer generated and almost certianly contains errors. This transcript is provided for information purposes only.EarningsCall, LLC makes no representation about the accuracy of the aforementioned transcript, and you are cautioned not to place undue reliance on the information provided by the transcript.
spk03: Good afternoon. My name is Mariama and I will be your conference operator today. At this time, I would like to welcome everyone to NVIDIA's financial results conference call. All lines have been placed on mute to prevent any background noise. After the speaker's remarks, there will be a question and answer session. If you would like to ask a question during this time, simply press star followed by the number one on your telephone keypad. If you would like to withdraw your question, press the pound or hash key. Thank you. I will now turn the call over to Simona Jankowski, NVIDIA Vice President of Investor Relations and Strategic Finance, to begin the conference.
spk01: Thank you. Good afternoon, everyone, and welcome to NVIDIA's conference call for the fourth quarter of fiscal 2021. With me on the call today from NVIDIA are Jensen Wang, President and Chief Executive Officer, and Colette Tress, Executive Vice President and Chief Financial Officer. I'd like to remind you that our call is being webcast live on NVIDIA's Investor Relations website. The webcast will be available through replay and through the conference call to discuss our financial results for the first quarter of fiscal 2022. The content of today's call is NVIDIA's property. It can be reproduced or transcribed without our prior written consent. During this call, we may make forward-looking statements based on current expectations. These are subject to a number of significant risks and uncertainties, and our actual results may differ materially. For a discussion of factors that could affect our future financial results and business, please refer to the disclosure in today's earnings release, our most recent Forms 10-K and 10-Q, and the report that we may file on Form 8-K with the Securities and Exchange Commission. All our statements are made as of today, February 24, 2021, based on information currently available to us. Except as required by law, we assume no obligation to update any such statements. During this call, we will discuss non-GAAP financial measures. You can find a reconciliation of these non-GAAP financial measures, GAAP financial measures, in our CFO commentary, which is posted on our website. With that, let me turn the call over to Collette.
spk04: Thanks, Simona. Q4 was another record quarter, with revenue exceeding $5 billion. and year-on-year growth accelerating to 61%. Full-year revenue was also a record at $16.7 billion, up 53%. Our gaming business has reached record revenue of $2.5 billion in Q4, up 10% sequentially, and up 67% from a year earlier. Full-year gaming revenue was a record at $7.8 billion, up 41%. Demand is incredible for our new GeForce RTX 30 series products based on the NVIDIA Ampere GPU architecture. In early December, we launched the GeForce RTX 3060 Ti, which joined the previously launched RTX 3090, 3080, and 3070. The entire 30 series lineup has been hard to keep in stock, and we exited Q4 with channel inventories even lower than when we started. Although we are increasing supply, channel inventories will likely remain low throughout Q1. GeForce RTX 30 series graphics cards were a holiday sensation. due not just to their amazing performance, but also to their rich features, including our second-generation RTX ray-tacing technology and DLSS, AI-powered performance accelerator, which massively boosts frame rates in graphically demanding titles. Three dozen games now support RTX, including the top battle royale game, Fortnite. the top role-playing game, Cyberpunk 2077, the top massively multiplayer online game, World of Warcraft, and the best-selling game of all time, Minecraft. RTX has clearly set the new standard in gaming. Building on this momentum, at CES in January, we introduced a wave of Ampere architecture gaming products, including our biggest ever laptop launch, powered by GeForce RTX 3060, 3070, and 3080 laptop GPUs, and with our third generation Max-Q technology. These new, thin, and lightweight gaming laptops increase performance and energy efficiency by up to 2x from the prior generation. RTX 3060 laptops start $999 and are faster than the previous generation laptops, which sold for $2,500. The incredible performance, design, and price points of these new laptops will delight the growing universe of gamers and creators, as well as students and professionals. The gaming laptop market has grown sevenfold in the past seven years, and momentum is building. With top OEMs bringing to market a record 70-plus laptop models based on the GeForce RTX 30 series. GeForce laptops as a whole are the fastest growing and one of the largest gaming platforms. Also at CES, we announced the GeForce RTX 3060 GPU, priced at $329, extending the 30 series desktop liner further into the mainstream. We expect strong demand when it launches this Friday, as 60 class GPUs have traditionally been our most popular products. Starting with the 3060, we're taking an important step to maximize the supply of GeForce GPUs for gamers. Users are constantly discovering new applications for our powerful programmable GPUs, and cryptocurrency mining is one of them. With rising Ethereum prices, there are indications that miners are behind GPUs. We would like GeForce GPUs to end up with gamers. So we have created a new special software drivers that will detect the Ethereum mining algorithm, cutting in half the mining efficiency of the GeForce RTX 3060. We suspect the significant increase in the Ethereum network hash rate observed over the past few months was driven by a combination of previously installed mining capacity that was reactivated, as well as new sales of GPUs and ASICs. Since our GPUs are sold to graphics card manufacturers and then on to distribution, we don't have the ability to accurately track or quantify their end use. Analyst estimates suggest that crypto mining contributed 100 to 300 million to our Q4 revenue, or a relatively small portion of our gaining revenue in Q4. Cryptocurrencies have recently started to be accepted by companies and financial institutions and show increased signs of staying power. To address industrial Ethereum mining demand, last week we announced a new line of NVIDIA CMPs, or crypto mining processors. Shipments will start in March. CMPs lack display outputs and have other optimizations that improve crypto mining power efficiency. CMP products will let us gain some visibility into the contribution of crypto mining to our overall revenue. For Q1, we estimate that CMP will contribute approximately $15 million. We plan to sell these products to industrial miners. We will quantify their contribution each quarter for transparency. Over the past year, it has become clear that we've entered a new era in which gaming is an integral part of global culture. The number of concurrent users on Steam has more than doubled since 2018 and continues to hit new records. In 2020 alone, more than 100 billion hours of gaming content was seen on YouTube, and half a billion people watched esports. Increasingly, we aren't just gaming. We're also watching sports, attending concerts, creating content, and connecting with our friends in virtual environments. Additionally, we are excited about the new experiences like VR. Significantly more content is now available, including arguably the first VR killer app, Beat Saber. There's now almost two million VR users on Steam. And with these powerful structural shifts, we expect our gaming business to remain on a robust growth trajectory. The GeForce RTX 30 series GPUs have kicked off a powerful upgrade cycle, and we estimate only around 15% of GeForce gamers own an RTX-class GPU, which is needed to experience the beautiful ray-traced graphics of modern games. Moreover, the universe of gamers is rapidly expanding, and the reach of GeForce has extended beyond gamers to some 45 million creators. In addition, gaming revenue continues to benefit from a favorable mix shift as gamers and creators keep moving to higher-end GPUs. We expect another great year for GeForce. Earlier this month, we celebrated the one-year anniversary of the GeForce Now cloud gaming platform, which is now over 6 million numbers strong. GeForce Now offers 800 PCs from over 300 publishers, more than any other cloud gaming service, including 80 of the most played free-to-play games. Starting with support for Windows PCs, Macs, and Android devices, we added support in recent months to Chromebooks, iPhones, and iPads. GFN has grown globally with more than 65 countries on our service and more added regularly by our GeForce Now Alliance partners. Moving to ProVis, Q4 revenue was at $307 million, up 30% sequentially and down 10% year on year and ahead of our expectations. Full year revenue was 1.1 billion, was down 13%. Strong sequential growth was driven primarily by a recovery in desktop workstations as some customers returned to the office and enterprises resumed purchases that had been deferred by the pandemic. Notebook GPUs grew sequentially to a record as enterprises continued to support remote workforce initiatives. Looking ahead, the reopening of businesses will benefit desktop workstations, but longer-term workforce trends will likely shift or mix to notebook GPUs and cloud offerings. Healthcare was a standout vertical in the quarter, with significant orders from GE, Siemens, and Oxford Nanopore Technologies. Public sector and automotive also sowed strength. Omniverse, our real-time 3D collaboration and simulation platform, is now in open beta. Over 500 creators and professionals have tested Omniverse through our early access program. Omniverse is one of our most important and exciting platforms. We are delighted by its initial acceptance and look forward to sharing more details on its long-term growth opportunity in the coming months. Moving to automotive. Q4 revenue was $145 million, up 16% sequentially and down 11% year-on-year. Full-year revenue of $536 million declined 23%. Sequential growth was driven by continued recovering the global automotive production volumes and growth in AI cockpit revenue. The year-on-year decline reflects the expected ramp-down of legacy entertainment. NVIDIA has emerged as the industry's leading end-to-end full-stack technology provider for self-driving and AI-enabled vehicles. Oren, the SOC that self-driving platform is built on, delivers an unrivaled 254 trillions of operations per second of performance on industry-leading power efficiency, helping to revolutionize the transportation industry. Our technology leadership is driven a robust, rapidly growing set of opportunities. We have great momentum with an expanding list of electric vehicle OEMs, including NIO, SAIC, Viato, and XPeng, which are all using the NVIDIA Drive platform to power their next generation of vehicles. We look forward to growing with them as they continue to scale. Our software-defined platforms is the only solution that spans from the data center for training deep neural nets and running physically accurate simulations to a full stack in car solutions, scaling from ADAS to level five fully autonomous functionality. Autonomous vehicle companies are harnessing this technology. Zoox recently unveiled its level five bi-directional robo taxi powered by NVIDIA. Enride launched its next-generation cabless autonomous truck using NVIDIA DRIVOREN. And earlier this year, Mercedes announced a 56-inch wide MBUX hyperscreen powered by NVIDIA AI cockpit technology. This win builds on our momentum with Mercedes' first-generation MBUX system, which is now in 1.8 million cars. We are in the early innings of a significant opportunity. We have built a multi-billion dollar design wind pipeline for our self-driving AI cockpit solutions, which will drive a material inflection in revenue over the next few years. Our transformational partnership with Mercedes, announced last June, demonstrates the power of our evolving business model as we expand our addressable market and layer in software revenue. We are exceptionally well-positioned to capitalize on the significant opportunity that lies ahead. Moving to data center. Revenue was $1.9 billion, which exceeded our expectations, was comparable to last quarter, and up 97% from the year-ago period, which did not include Mellanox. Data center compute revenue was up 45% year-on-year. Full-year data center revenue rose 125% to a record $6.7 billion, including almost 70% growth from data center compute. From a sequential perspective, the data center computes stronger than expected double-digit growth more than offset the anticipated decline in Mellanox revenue, which included a large non-recurring network sale to a single OEM in Q3. Compute growth was led by vertical industries, where OEM partners continued ramping up their A100-based servers, and our own DGX system sales were strong. Vertical industries were well over 50% of data center revenue across compute and networking, with particular strength in supercomputing, financial services, higher education, and consumer internet verticals. Additionally, hyperscale customers continue to deploy the A100, driving both sequential growth and exceptionally strong year-on-year growth in data center compute. The A100 has been adopted by all major cloud customers globally and is being deployed by hyperscale customers for internal workloads. Still, we are in the early stages of adoption and expect continued growth this year. of the A100 has been smoother and accomplished by better visibility than prior generation. Its universal AI training and inference capabilities, as well as support for a wider set of applications and outstanding performance are driving high customer utilization, a clear sign of the A100's value. Turning to Mellanox, we are seeing continued strong traction and robust momentum across our customer sets. Its revenue was up over 30% from Mellanox's Q4 revenue in calendar 2019, when it was still a standalone company. Year-on-year growth in the quarter was led by hyperscale and large customer consumer internet customers, which grew over 60% from last year, with several contributing record revenues. Consistent with our outlook, Mellanox had a sequential decline impacted by a non-recurring sales to a China OEM in Q3. We expect to return to sequential growth in Q1, driven by strong demand for our high-speed networking products, including the ramp of ConnectX adapters with CSPs and all major server OEMs in their upcoming refresh. We also see strong momentum in high-performance computing with HDR InfiniFAM products. For example, we won six of the seven supercomputers awarded over the past few months by EuroHPC. Starting next quarter, we will continue to provide color and networking as part of the data center market platform, but we will no longer break out Mellanox revenue separately. Looking forward, we are incredibly excited about the opportunities in data center. Accelerated computing is not only delivering super more laws gains in performance, but is also an energy efficient and cost effective method of computing. And virtually every industry is adopting technology with greater urgency as companies adapt to the new world of more distributed workers and customers. As industries embark on this journey, they are also increasingly focused on combating climate change. To that end, the A100 performs AI computations with 1 20th the power consumption of CPUs. It powers our Selene supercomputer, which is number one on the green 500 list of the world's most efficient supercomputers. Indeed, NVIDIA's powered machines recently captured 25 of the top 30 spots on the green 500 list. Accelerated computing is not only serving the exponential growth in demand for compute, it can also help bend the power consumption curve. With accelerated computing, NVIDIA is pioneering a path forward the computing industry. Before I move to the P&L and Outlook, let me give you an update on our proposed acquisition of Arm. In September, we announced plans to acquire Arm from SoftBank Group in a transaction that will create the premier computing company for the age of AI. At that time, we said it would take approximately 18 months to secure regulatory approvals in the US, the UK, the EU, China, and other jurisdictions. Thorough reviews are typical with a deal of this size. This process is moving forward as expected. We are in constructive dialogue with the relevant authorities and are confident that regulators will see the benefits to the entire tech ecosystem. As we have said, this combination will spur competition. Together, ARM and NVIDIA will provide greater choice to the data center ecosystem, a compelling alternative CPU architecture for the market, and further enhance ARM's offering in mobile and embedded. Our intention is to increase investment in ARM's existing roadmap, adding resources to stimulate growth in new markets. We love and intend to maintain ARM's open licensing model, a commitment guaranteed both by long-term legally binding contracts, as well as our own interest in ensuring this investment is a profitable one for us. We are on the cusp of a new age in which AI fuels industries ranging from healthcare to scientific research to the environment. With this transaction, our vision is to boost arms potential so it can thrive in this new era and grow into promising new markets. Moving to the rest of the P&L. Q4 gap gross margins were 63.1% and non-gap gross margins were 65.5%. Gap gross margins declined year on year due to amortization of developed technology acquired from Mellanox, partially offset by product mix. The sequential increase was due to higher margins for gaming GPUs and lower IP-related costs, partially offset by lower margin mix in our data center portfolio. Non-GAAP gross margins increased by 10 basis points year on year and was flat sequentially in line with our expectations. Q4 GAAP EPS was 2.31, up 51% from a year earlier. Non-GAAP EPS was $3.10, up 64% from a year ago. Q4 cash from operations was a record, 2.07 billion. With that, let me turn to the outlook for the first quarter of fiscal 2022. Revenue is expected to be 5.3 billion, plus or minus 2%, with most of the sequential growth driven by gaming. Gap and non-gap gross margins are expected to be 63.8% and 66%, respectively, plus or minus 50 basis points. GAAP and non-GAAP operating expenses are expected to be approximately $1.67 billion and $1.2 billion, respectively. For the full year, we expect to grow non-GAAP OpEx in the mid-20% range. GAAP and non-GAAP other income and expenses are both expected to be an expense of approximately $50 million. Gap and non-gap tax rates are both expected to be 10%, plus or minus 1%, excluding discrete items. Capital expenditures are expected to be approximately $300 million to $325 million. Further financial details are included in the CFL commentary and other information on our IR website. In closing, let me highlight upcoming events for the financial community. We will be virtually attending the Raymond James Institutional Investors Conference on March 1st, the Morgan Stanley Technology, Media, and Telecom Conference on March 3rd, and the Arate Virtual Semis Conference on March 3rd. In addition, we will be hosting a virtual investor day on Monday, April 12th, following the live stream of Jensen's opening keynote at our GPU technology conference. Our earnings call to discuss our first quarter and full, our first quarter is scheduled for Wednesday, May 26th. We will now open the call for questions. Operator, would you please talk for questions? Thank you.
spk03: Yes, at this time, I would like to remind everyone in order to ask a question, press star, then the number one on your telephone keypad. We'll pause for just a moment to compile the Q&A roster. Your first question comes from the line of CJ Muse with Evercore ISI. Your line is open.
spk09: Good afternoon. Thank you for taking the question. I guess Jensen higher level question for you on the enterprise side. You're now a couple quarters into the ramp of a 100 and curious if you can speak to whether you've seen any surprises here, any areas of specific strength worth calling out and any changes to how you're thinking about the size of this opportunity.
spk07: Yeah, thanks a lot. As you know, a 100 is a very different type of GPU. This is our first universal computing GPU. It's great at high-performance computing. It's great at data analytics. It's great at training. And also, for our highest-end GPU, it's also the first time that it's incredible for inference. It's some 20 times faster than previous generation, and it introduces some really, really exciting new computational formats like TF32, TensorFlow32 for training, and M8 with multi-instance GPU turning our GPU, one GPU, into a whole bunch of smaller GPUs, autonomous GPUs, to improve performance and reducing latency. And so the capability is really quite exciting. We're seeing strength in hyperscalers as they continue to accelerate their adoption of AI. Some of the new applications we've spoken about a couple times before, the transition to deep learning for conversational AI, speech recognition to natural language understanding, all the way to speech synthesis is now based on AI, based on deep learning. The other area that's growing incredibly fast is deep learning recommender models. Just about everything that you do on the Internet is based on recommenders. There are hundreds of different recommenders out there, whether you're shopping or recommending music or recommending news or recommending search. And so all of the recommending ads. And so all of these different types of applications are driving that. For the first time, we saw our industrial application, industrial data center growing to be larger than hyperscale. And we're seeing industrial applications across scientific computing where simulation-based approaches are now being fused with AI approaches for weather simulation, genomics, molecular dynamics simulation, quantum chemistry, even simulating quantum computing, which is one of the really exciting areas. We're seeing AI being deployed for big data analytics, RAPIDS, which is NVIDIA's created open source platform for data analytics. Spark 3.0, which NVIDIA really led and is now GPU-accelerated. So now you could have big data in the cloud while doing big data analytics in the cloud on all of the CSP platforms. You could receive a lot of excitement around financial services. financial services and consumer internet services are all really growing nicely. And so A100 adoption is just starting. I mean, we're going to see several, you know, a couple of years of continued growth yet ahead of us, you know, as AI gets adopted in clouds and industries.
spk03: Your next question comes from the line of Vivek Arya for Securities. Your line is open.
spk00: Thanks for taking my question. Just a clarification and then a question for Jensen. On the clarification, Colette, I was hoping if you could give a little more color around Q1. Do you still expect the data center to grow sequentially in Q1? I know you said that most of the growth will come from gaming, but any color on the data center would be useful. And then, Jensen, the question for you is, in your press release, you used the phrase AI driving the smartphone moment for every industry. Could you help us quantify what that means? And where I'm going with that is, is there a number in terms of what percentage of servers are shipping today with your accelerators? And where can that ratio go over time? Is that a fair way of looking at the adoption of your technology and AI?
spk04: So thank you, Vivek, your question regarding the guidance as we lead into Q1. We had indicated that, yes, a good percentage of our growth between Q4 and Q1 rolls down from gaming. But we also do expect data center to grow. Most of our sequential growth coming from gaming. But keep in mind, we also expect all of our market platforms will likely be able to grow quarter over quarter.
spk07: We're entering the third phase of AI. The first phase of AI was when we invented the computing platforms, the new chips, the new systems, the new system software, the new middleware, the new way of working, the new way of developing software, which the industry, the world is now starting to call MLOps. The way that software is developed and the way that it's deployed is completely different than the past. In fact, I heard a great term, software 2.0, and it makes a lot of sense. It's a computer that is writing software. The way that you develop software is completely different. The way you compute is different. And that was our first phase, and that started in a journey that was some eight, nine years ago now. The second phase was the adoption of using this in an industrial way for clouds. And we saw it revolutionize new services, whether it's speech-oriented services or search-oriented services, recommender services, the way you shop, the way you use the internet is completely different today. And so that's really the second phase. And those two phases are still continuing to grow. And you're still seeing The growth associated with that. The third phase is the industrialization of AI. And some of the great examples, when I say kind of a smartphone moment, I meant that it's a device with AI, it's autonomous, and it's connected to a cloud service, and it's continuously learning. So some of the exciting examples that I saw that I've seen and we're working with companies all over the world, we have some 7,000 AI startups that we're working with, and almost all of them are developing something like this. And large industrial companies, whether it's John Deere or Walmart, they're all developing applications kind of like this. And basically, it's an autonomous system, autonomous machine, In our case, it's called Jetson. It's a robotics machine. If that robotics machine is a car, then it's called Drive. And it's running an autonomous, an AI application on top, an AI skill on top. And it could be moving things around. It could be picking and placing. It could be just watching a warehouse and monitoring traffic and keeping traffic flow going, it could be connected to a car. And whenever the fleet of cars needs to be retrained because of a new circumstance that was discovered, the cloud service would do the relearning and then would deploy it into all of the autonomous devices. And so in the future, we're seeing that these industries, whether you're in retail or logistics or transportation or you know, farming, ag tech, you know, to lawnmowers, consumer lawnmowers. They're not going to just be products that you buy and use from that point forward, but it'll likely be a connected device with an AI service that runs on top of it. And so these industries are so excited about it because it gives them an opportunity to change the way that they interact with their customers. Rather than selling something once They sell something and provide a service that's on top of it. And they can stay engaged with the customers. The customers could get a product that's improving all of the time, just like your smartphone. And that's kind of the reason why I've been calling it the smartphone moment for all these industries. And we saw what happened to the smartphone revolution. And then we saw what happened to the smart microphone, the smart speaker revolution. you're going to see smart lawnmowers, smart tractors, smart air conditioners, smart elevators, smart buildings, smart warehouses, robotic retail stores. The entire retail store is like a robot. And they will all have autonomous capability. They'll all be driven by AI. And so what's new for the industry, therefore, is that all of the enterprises in the world used to have computers for IT to facilitate, to host their employees and their supply chain. But in the future, all of these industries, whether you're in medical imaging or, you know, or in lawnmowers, you're going to have data centers that are hosting your products, just like the CSPs. And so that's a brand new industry. And we have a platform that we call EGX, which is the 5G edge AI systems. And we have the autonomous systems we call AGX, which is what goes into jet planes and drives. And between those two systems and the software stack that we have on top of it, we're in a great position to help these industries one at a time transform their business model from an object-oriented business model, a thing-based business model, to a connected device business model.
spk03: Your next question comes from the line of Stacy Rosgon with Bernstein Research. Your line is open.
spk06: Hi, guys. Thanks for taking my question. First, I don't want to be pedantic, I suppose, but I guess on the Q1 guide, you're saying that gaming is the majority of the growth. Was that an absolute statement or was that a percentage statement? Can you give us some idea of how you'd sort of rank the sequential percentage growth of, say, you know, gaming versus data center versus other, especially since it sounds like you've got 50 million users? in crypto-specific stuff, but it'll go into the other. And then I guess just briefly, could you give us some indication of where your supply situation and lead times are on your Ampere parts within data center? I think you said last quarter they were many months on six months plus. Are they still looking like that? And is that sort of the limiting factor at this point in terms of what you can actually ship on the compute side of data center? Caleb will take one, and I'll take the other.
spk04: Sure. Let me start off, Stacey, in terms of our guidance for Q1. As you know, we're still in the early innings of our Ampere architecture, our Ampere architecture as it relates to gaming as well as what it relates to data center. As we articulated in our call, we have been really seeing continued uplift folks adoption of a 100 and it's going quite smoothly than what we had seen in prior overall versions so when we think about our guidance for q1 there's many different types of conclusions that will happen at the end of the quarter in terms of what we said but all of our platforms can grow but the majority of the majority of the growth in from Q4 to Q1 will likely be gaming.
spk07: You asked a question about lead time. Our company, at the company level, we're supply constrained. Our demand is greater than our supply. And however, for a data center, so long as the customers work closely with us and we do a good job planning between our companies, there shouldn't be a supply issue for data centers. We just have to do a good job planning. And we have direct relationships with each one of the world CSPs. And we have direct relationships with all the OEMs. And we could do excellent planning between us. We shouldn't have a supply, we shouldn't be supply constrained there. But at the company level, we're supply constrained. Demand is greater than supply. And we have enough supply We have enough supply. We usually have enough supply to achieve better than the outlook. And we had that situation in Q4. We expect that situation in Q1. And we have enough supply to grow through the year. But supply is constrained and demand is really, really great. And so We just have to do a really good job planning. Meanwhile, one of the things that really came through for us is we have the world's best operations team. Our company really has an amazing operations team. We build the most complex products in the world, the most complex chips, the most complex packages, the most complex systems. And during Q4, they improved our cycle time. And during Q1, I'm expecting them to improve our cycle time again. And we really are blessed to have such an amazing operations team. And so during these times, it really comes in handy. But overall, at the company level, we expect demand to be greater than supply. We have enough supply to do better than the outlook. And we have enough supply to grow each quarter throughout the year.
spk03: Our next question comes from the line of Timothy Arcuri with UBS. Your line is open.
spk05: Hi, thanks. I had a question on crypto. I guess, Jensen, I know that the CNP stuff and the software driver stuff that you're doing for the 3060, that's going to help a lot. But I think that there's like four or five of the big currencies are going to move, or at least they're moving or they're on a path to move from a proof of work to a proof of stake. which is going to be a lot less compute intensive. So I guess the question that I get a lot is, how do you assess the degree to which that drives GPUs back into the secondary market? Is there any way that you can get kind of a handle on that?
spk07: Thanks. Yeah, if you look at the recent hash rates, first of all, the transition is going to take some time. It can't happen overnight. And people have to build trust in the new versions. And so it'll take a little bit of time. But I hope it does. I hope that people use proof of stake over time and a little bit of these questions don't have to be answered. However, I don't have that much optimism either that it will be all proof of stake. I think that... Proof-of-work is a very legitimate way of securing the currency in the beginning. While any currency is building its reputation, it's going to take something like a proof-of-work to do so. And so I think proof-of-work is going to be around for a bit. We developed CMP for this very reason, just so that there are different versions. We have different versions of our products for gaming. for professional visualization, for high-performance computing, for deep learning. It stands to reason we have the ability to do a different version for CMP. And we can sell it directly. The way that we go to market would be to go directly to the industrial miners. And it's a great benefit to them so that they don't have to chase around spot markets. It's a great benefit to the gamers because they want to game. And the gaming demand is just incredible. It's off the charts. And so I think this is going to be really beneficial to everybody. The recent hash rate growth was really a result of several dynamics. The first dynamic is the install base. Most people thought that once they're mining, the GPUs come back into the aftermarket. A small part does that. Some people do that. But the vast majority don't keep them. And the reason for that is because, obviously, they believe in Ethereum. And so they're industrial miners. That's what they do. And so they keep it around for when the profitability returns and they could kickstart their mining gear. That's what we saw in the latter part of last year. We saw the hash rate starting to grow. Most of that was a result of the installed miners reactivating their equipment. It wasn't until earlier this year that we started to see demand in our own GPUs. And when that starts to happen, there are several different dynamics. The primary source these days come from powerful ASICs. And then there's some that comes from our GPU and other GPUs in the marketplace. And so I think that this is going to be a part of our business. it will grow extremely large no matter what happens. And the reason for that is because when it starts to grow large, more ASICs comes in the market, which kind of mutes it. And when the market becomes smaller, it's harder for ASICs to sustain the R&D. And so And so the spot miners, industrial miners come back and then we'll create CNPs. And so we expect it to be kind of a, to be a small part of our business as we go forward. Now, one of the important things is to realize that in the near term, because we're in the beginning parts of our Ampere ramp, only two quarters into it, into a multi-year cycle, This is also the first time that we've completely changed computer graphics. RTS, using ray tracing, is completely different than rasterization. And so this is a fundamental change in the way we do computer graphics, and the results have been spectacular. There are some 200 million install base and desktops and 50 million in laptop, and the vast majority of them we've only upgraded approximately, I think it's something like 15%. of the install base has been upgraded to RTX. And so there's a giant install base and the install base is growing that we need to upgrade to the next generation of computer graphics.
spk03: Your next question comes from the line of John Pitzer with Credit Suisse. Your line is open.
spk08: Yeah, guys, thanks for letting me ask the question. I want to go back to data center. You've been very kind over the last couple quarters to call out Mellanox, both when it was a positive driver and when it was a headwind. I'm kind of curious, when you look into the fiscal first quarter, is there anything of distinction to mention around Mellanox versus core data center? And I guess as a follow-on, you know, the key metric that a lot of investors are looking at is when does the core data center business year-over-year growth start to re-accelerate. And some of that is just simple math. You're just comping very hard compares from last year. But, Jensen, how would you think about data center year-over-year growth in the context of a reopening trade or of any sort of new applications out there? I mean, what helped last time around was the move to natural language AI. Is there another big sort of AI application we should be thinking about as we think about data center growth re-accelerating?
spk07: Yeah, we're expecting – Mellanox was down this last quarter, and our compute business grew double-digit and offset – more than offset the decline in Mellanox. We expect Q1 to be a growth quarter for Mellanox, and we expect this coming year to be quite an exciting year for Mellanox. The business is growing in Ethernet, it's growing for CSPs, it's growing in FinBand for high-performance computing, and the switches are growing. Switch business grew 50% year-over-year. And so we're seeing really terrific growth there. One of the new initiatives, and we're going to see success towards the second half because the number of adoptions and number of engagements is so great, is our new Bluefield DPUs. It's used for virtualization for hyperscalers. It's also used for security. As you know quite well, the future of computing is cloud, and it's multi-timing cloud. And there's no VPN front door to the cloud. You've got millions of people who are using every aspect of your computing, so you need to have distributed firewalls, and you can't have it just in one place. The intense focus of security across all of the data centers around the world is really creating a great condition for Bluefield, which is really perfect for them. And so I expect our Mellanox networking business to grow very nicely this year. And we expect Q1 to be a great growth quarter for compute as well as Mellanox. The killer, the great driving application for AI are several. Last year, you're absolutely right that it was natural language understanding. The transformer model and what is the core of BERT and other versions like that really made it possible for us to enable all kinds of new applications. So you're going to see natural language understanding do text completion and It's going to be integrated. I think it was just announced today that it was going to be integrated into Microsoft Word. We've been working with them on that for some time. And so there's some really exciting applications like that. But the new ones that emerged recently are deep learning-based conversational AI, where the ASR, the speech recognition, as well as the speech synthesis are now based on deep learning. It wasn't before. And they were based on models that ran on CPUs. But now with these deep learning models, the accuracy is much, much higher. And it has the ability to also mimic your voice and be a lot more natural. And so the ability to, these models are much more complex and much larger. The other big, huge driver is recommenders. This is something really worthwhile to take a look at. It's called deep learning recommender models. And recommenders have historically, whether it's for shopping or personalizing websites or personalizing your store or recommending your basket, recommending your music, historically it's been used as traditional machine learning algorithms. But because of the accuracy and just the extraordinary economic impact that comes from an incremental 1% in accuracy, for most of the world's large internet businesses, people are moving very rapidly to deep learning-based models. And these models are gigantic. They're utterly gigantic. And this is an area that is really driving hyperfinance computing. So I expect us to see a lot of momentum. And the last one is the one that I just spoke about, which has to do with industrial 5G and edge IoT type of applications for all of the different industries, whether it's retail or logistics or transportation, agricultural or warehouses to factories. And so we're going to see AI and robotics in a very large number of applications and industries. And we're just seeing so much excitement there.
spk03: Your next question comes from the line of Aaron Rakers with Wells Fargo. Your line is open.
spk10: Yeah, thanks for taking the questions. I wanted to go back again on the data center business. You just mentioned, Jensen, the Bluefield 2 product, you know, poised to kind of ramp and materialize in the back half of the calendar year. How do you see that? You know, is it an attach rate? I think there's been discussions in the past about, you know, all servers could potentially over time incorporate this layer of acceleration, how quickly should we think about that ramp? And then the second question is, can you just at a high level talk about how CPU, how a CPU strategy, you know, you're thinking about that in the context of the broader data center market?
spk07: Sure. If I could just work backwards, I believe that every single data center node will be outfitted with a GPU someday. And that someday is probably, call it five years from now. And the fundamental driver of it is going to be security. Every single application in the data center and every single node in the data center has to be individually secured. Zero trust computing, zero or confidential computing and zero trust computing, these initiatives are going to cause every data center to have every single application and every single node be secured. Which means every one of those computers have to have a control plane that is isolated from the application plane. And all the applications cannot share the same resources because that application could be malware, that application could be an intruder, that no application could have access to the control room. And yet today, the software-defined data centers, the software-defined networking, software-defined storage, all of the security agents are running in the same processors as the applications, and that hasn't changed. You see the cloud CSPs in the world moving in this direction. Every single data center will have to moving in this direction. So every node will be a DPU process for the infrastructure. You're essentially going to see the data center infrastructure be offloaded from the application plane. And it will be something like a root field. So I think this is our next multi-billion dollar opportunity. We support every CPU in the world. We're the only accelerated computing platform that accelerates every CPU. Ironically, the only CPU we don't accelerate for AI is ARM. But we want to change that. ARM has such an exciting future because the nature of their business model and the nature of their architecture is perfect for the future of hyperscalers and data centers. You want the most energy efficiency in every single data center because every data center is power constrained. We are going to be power constrained in every aspect of computing going forward. And so we would love to build around the ARM processor and invest in building a great ecosystem around it so that all of the world's peripherals and all the world's applications to work on any one of the CPUs that we know today. And we're going to start with high-performance computing and start with AI and all the areas that we have a lot of expertise in, build out that platform. So you're starting to see one industry leader after another embrace ARM. And I think that's terrific. But now we've got to... energize it with all of the ecosystem support. It can't just be vertical applications, but we want to create a broad and general ARM ecosystem.
spk03: Your next question comes from the line of Mark Lipakchus with Jefferies. Your line is open.
spk02: Hi. Thanks for taking my question. A question for Jensen, I think. Jensen, if you look at the past computing eras, typically it's one ecosystem that captures 80% of the value of that computing era. In many frames, there's IBM. In many computers, there's DEC, PCs, Wintel, cell phones, Nokia, and then Apple. So if you don't get the ecosystem right, then you're splitting 20% of the market with a with a handful of players. So in this next era of computing, parallel processing or AI, I think you've articulated the most compelling architectural vision of the data center of the future with data center scale computing devices, with CPUs, GPUs, DPUs integrated in the same box, serving all workloads, I imagine, in a virtualized environment. Can you help us understand where's the market in embracing that vision and where's NVIDIA in building out the ecosystem for that data center scale computing vision? And maybe as part of that, to what extent is CUDA the kernel for that ecosystem? Thank you.
spk07: Yeah, I think we've done a great job building out the platforms for several ecosystems around the world. And the domains that we do incredibly well at are the domains that have to do with accelerated computing. We pioneered this approach, and we brought it to high-performance computing at first, and we accelerated scientific computing and we democratized supercomputing for all researchers. Anybody who wants to have a supercomputer now can. And computing will simply not be the obstacle to somebody's discovery. We did the same for artificial intelligence. We did the same for visualization. We expanded the reach of gaming tremendously. GeForce today is the largest gaming platform. It's the single largest body of computers that are used for gaming. And in each case, we expanded the market tremendously. We would like to do the same for data center scale computing as it applies to virtualizing these applications. These applications are also in the process. They've historically required dedicated systems, but they're moving into a virtualized data center environment. And we are best at doing that. They run on our platform today. We have the ability to virtualize it and put it into the data center and make it remotely available. And so these applications, these domains are some of the most important domains in the world. And so we're in the process of doing that. By doing so and making our architecture available to CSPs and OEMs, we could create this accelerated computing platform available to everybody. And so you're seeing our journey doing that. First, creating and architecting this platform, and then putting it literally into every single data center in the world. The next step of our journey is the phase three of AI. And it has to do with turning every endpoint into a data center, whether it's a 5G tower, a warehouse, a retail store, a self-driving car, a self-driving truck. These are going to be – they're all going to be essentially autonomous data centers. And they're going to run AI, but they're going to run a lot more. They're going to do security in real time. Its networking is going to be incredible. It's going to run software with 5G and GPU-accelerated 5G we call Arial. And so these platforms are going to become data centers. They'll be secure. The software is protected, and you can't tamper with it. If you tamper with it, the course won't run. And so the capability of these clouds will move all the way out to the edge. And we're in the best position to be able to do that. So I think in this new world of post-Moores law, post-Bennard scamming, In this new world where AI is software that writes software, in this new world where data centers are going to be literally everywhere and they're unprotected, there's no giant building with a whole bunch of people that secure it, and in this new world where software is going to enable this autonomous future, I think we are perfectly positioned for it.
spk03: This is all the time we have for Q&A today. I will now turn the call back to CEO Jensen Wang.
spk07: Thanks for joining us today. Q4 capped a truly breakout year for NVIDIA. The two biggest engines of our business, gaming and data center, posted powerful growth. Gaming has become the world's largest media and entertainment industry and will grow to be much larger. And again, gamers will create, will play, they'll learn, they'll connect, The medium of gaming can host any type of game and eventually evolve into countless metaverses, some for play, some for work. Gaming is simultaneously a great technology and a great business driver for our company. This year, we also closed our Mellanox acquisition and successfully united the amazing talent of our companies. Combined, we possess deep expertise in all aspects of computing and networking, to drive the architecture of modern data centers. Cloud computing and hyperscalers have transformed the data center into the new unit of computing. Chips and servers are just elements of the data center scale computers now. With our expertise in AI computing, full-stack accelerated computing, our deep network to computing expertise, and cloud-to-edge platforms, NVIDIA is helping to drive a great computer industry transformation. In our planned acquisition of ARM, the world's most popular and energy-efficient CPU company, will help position NVIDIA to lead in the age of AI. This year was extraordinary. The pandemic will pass, but the world has been changed forever. Technology adoption is accelerating across every industry. Companies and products need to be more remote and autonomous. This will drive data centers, AI, and robotics. This underlies the accelerated adoption of NVIDIA's technology. The urgency to digitize, automate, and accelerate innovation has never been higher. We are ready. We look forward to updating you on our progress next quarter. Thanks a lot.
spk03: This concludes today's conference call. You may now disconnect.
Disclaimer