NVIDIA Q1 2023 Earnings Call Transcript

Key Takeaways

  • Record data center revenue of $3.8 billion, up 15% sequentially and 83% year-over-year, driven by broad A100 GPU adoption for training and inference across hyperscale and vertical industries.
  • Gaming revenue hit a Q1 record of $6 billion (up 6% sequentially, 31% YoY) on the RTX 30 Series cycle, but NVIDIA expects a sequential gaming revenue decline in Q2 amid macro headwinds and channel normalization.
  • Unveiled the H100 Hopper GPU, delivering an order-of-magnitude performance leap over A100, and confirmed the upcoming Grace CPU for AI-optimized data center “factories” slated for launch in H1 FY23.
  • Automotive revenue grew 10% sequentially to $138 million as the Drive Orin SoC enters production with 35+ design wins and an automotive pipeline exceeding $11 billion over the next six years.
  • Q2 guidance calls for revenue of $8.1 billion ±2%, non-GAAP gross margin of ~67.1% ±50 bps, and expects strong data center and auto growth to more than offset gaming headwinds including a ~$500 million impact from Russia and China.
AI Generated. May Contain Errors.
Earnings Conference Call
NVIDIA Q1 2023
00:00 / 00:00

There are 14 speakers on the call.

Operator

Good afternoon. My name is David, and I'll be your conference operator today. At this time, I'd like to welcome everyone to NVIDIA's First Quarter Earnings Call. Today's conference is being recorded. All lines have been placed on mute to prevent any background noise.

Operator

After the speakers' remarks, there will be a question and answer session. Thank you, Simona Jankowski. You may begin your conference.

Speaker 1

Thank you. Good afternoon, everyone, conference call for the Q1 of fiscal 2023. With me today from NVIDIA are Jensen Huang, President and Chief Executive Officer Anne Colette Kress, Executive Vice President and Chief Financial Officer. I'd like to remind you that our call is being web webcast live on NVIDIA's Investor Relations website. The webcast will be available for replay until the conference call to discuss our financial results Q2 of fiscal 2023.

Speaker 1

The content of today's call is NVIDIA's property. It can be reproduced or transcribed without our prior written consent. During this call, presentation. During this call, we may make forward looking statements based on current expectations. Financial results and business.

Speaker 1

Please refer to the disclosure in today's earnings release, our most recent Forms 10 ks and 10 Q and the reports that we may file on Form 8 ks with the Securities and Exchange Commission. All our statements are made as of today, May 25, 2022, Investor Relations based on information currently available to us. Except as required by law, we assume no obligation to update any such statements. Financial measures. During this call, we will discuss non GAAP financial measures.

Speaker 1

You can find a reconciliation of these non GAAP financial measures financial measures in our CFO commentary, which is posted on our website. With that, let me turn the call over to Colette.

Speaker 2

Thanks, Simona. We delivered a strong quarter, driven by record revenue in both data center and gaming, strong fundamentals and execution against a challenging macro backdrop. Total revenue of $8,300,000,000 was a record, $6,000,000,000 rose 6% sequentially and 31% year on year, powered by the GeForce RTX 30 Series product cycle. Since launching in the fall of 2020, the RTX 30 Series has been our best gaming product cycle ever. The gaming industry has grown tremendously with 100,000,000 new PC gamers added in the past 2 years according to Newzoo.

Speaker 2

And NVIDIA RTX has set new standard for the industry with demand from both first time GPU buyers as well as those upgrading their PCs to experience the 250 plus RTX optimized games and apps, double from last year. We estimate that almost a third of the GeForce gaming GPU installed base is now on RTX. RTX has brought tremendous energy into the gaming world and has helped drive a sustained expansion in our higher end platforms and installed base Q1 results. And demand in Americas remains strong. However, we started seeing softness in parts of Europe Q1 results.

Speaker 2

Thank you, operator. Thank you, operator. Thank you, operator. Thank you, operator. Thank you, operator.

Speaker 2

Thank you, operator. Thank you, operator. Thank you, operator. Call. As we expect some ongoing impact, as we prepare for a new architectural transition later in the year, We are projecting gaming revenue to decline sequentially in Q2.

Speaker 2

Channel inventory has nearly normalized and we expect it to remain around these levels in Q2. The extent in which cryptocurrency mining contributed to gaming demand is difficult Q4. The reduced pace of increase in Ethereum network cash rate likely reflects lower mining activity on GPUs. We expect to diminishing contribution going forward. Laptop gaming revenue posted strong sequential and year on year growth, driven by the ramp of the NVIDIA RTX 30 Series lineup.

Speaker 2

With this year's spring refresh ahead of the upcoming back to school season, there are now over 180 laptop models Q1, up from 140 at this time last year. Driving this growth are not just gamers, but also the fast growing category of content creators from whom we offer dedicated NVIDIA Studio Drivers. We've also developed applications and tools to empower artists The creator economy is estimated at $100,000,000,000 and powered by 80,000,000 individual creators and broadcasters. We continued to build out our GeForce NOW cloud gaming service. Gamers can now access RTX 3080 cost streaming, new top tier offering with subscription plans of $19.99 a month.

Speaker 2

We added over 100 games to the GeForce NOW library, Bringing the total to over 1300 games. And last week, we launched Fortnite on GeForce NOW with touch controls for mobile devices, streaming through the Safari web browser on iOS and the GeForce NOW Android app. Moving to pro visualization. Q1 revenue was $622,000,000 was down sequentially 3% and up 67% from a year ago. Demand remains strong as enterprises continued to build out their employees' remote office infrastructure to support hybrid work.

Speaker 2

Sequential growth in the mobile workstation GPUs was offset by lower desktop revenue. Strong year on year growth was supported by the NVIDIA RTX AMP peer architecture product cycle. Top use cases include digital content creation at customers such as Sony Pictures, Animation and Medical Imaging at customers such as Medtronic. Software is being adopted by some of the world's largest companies. Amazon is using Omniverse to create digital twins to better optimize warehouse design and to train more intelligent robots.

Speaker 2

Kroger is using Omniverse to optimize store efficiency with digital twin store simulation. And PepsiCo is using Omniverse Digital Twins to improve the efficiency and environmental sustainability of its supply chain. Omniverse is also expanding our GPU sales pipeline, driving higher end and multiple GPU configurations. The Omniverse ecosystem continues to rapidly expand with 3rd party developers in the robotics, Industrial Automation, 3 d Design and Rendering Ecosystems Developing Connections to Omniverse. Moving to automotive.

Speaker 2

Q1 revenue of $138,000,000 increased 10% sequentially and declined 10% from the year ago quarter. Our Drive Orin SoC is now in production and kicks off a major product cycle with auto customers ramping in Q2 and beyond. Oren has great traction in the marketplace with over 35 customer wins from automakers, truck makers and robotaxi companies. Q1 BYD, China's largest EV maker and Lucid, an award winning EV pioneer, were the latest

Speaker 3

fleet to announce that they are

Speaker 2

building their next generation fleets on Drive Orin. Our automotive design win pipeline now exceeds 11,000,000,000 conference call over the next 6 years, up from $8,000,000,000 just a year ago. Moving to data center, record revenue of $3,800,000,000 grew 15% sequentially and accelerated to 83% growth year on year. Revenue from hyperscale and cloud computing customers more than doubled year on year, customer service provider. Customers remain supply constrained in their infrastructure needs and continue to add capacity as they try to keep pace with demand.

Speaker 2

Revenue from vertical industries growth grew a strong double digit percentage from last year. Top verticals driving growth this quarter include consumer Internet Companies, financial services and telecom. Overall, data center growth was driven primarily by strong adoption of our A100 GPU for both training and inference with large volume deployments by hyperscale customers and broadening adoption across the vertical industries. Top workloads includes recommender systems, conversational AI, large language models and cloud graphics. Networking revenue accelerated on strong broad based demand for our next generation 25, 50 100 Gig Ethernet Adapters.

Speaker 2

Customers are choosing NVIDIA's networking products for their leading performance and robust software functionality. In addition, networking revenue is benefiting from growing demand for DGX SuperPods and cross selling opportunities. Customers are increasingly combining our compute and networking products ability to build what are essentially modern AI factories with data as the raw material input and intelligence as the output. Our networking products are still supply constrained, though we expect continued improvement throughout the rest of the year. One of the biggest workloads driving adoption of NVIDIA AI is natural language processing, which has been revolutionized by transformer based models.

Speaker 2

Recent industry breakthroughs traced to transformers include large language models like GPT-three, NVIDIA Megamole Bart for drug discovery and DeepMind Alphafold for protein structure prediction. Transformers allow self supervised learning without the need for human labeled data. To do that, transformers use enormous training data sets and very large neuron networks, well into the hundreds of billions of parameters. To run these giant models without sacrificing low inference times, customers like Microsoft are increasingly deploying NVIDIA AI, including our NVIDIA Ampere architecture based GPUs and full software stack. In addition, we are seeing a rising wave of customer innovation using large language models At GTC, we announced our next generation data center GPU, the H100, based on the new upper architecture.

Speaker 2

Packed with 80,000,000,000 transistors, H100 is the world's largest most powerful accelerator offering an order of magnitude leap in performance over the A100. We believe H100 is hitting the market 2 largest scale AI workloads today. We are working with leading server makers and hyperscale customers Q1 2019 as well as the new DGX H100 AI computing system will ramp in volume late in the calendar year. Building on the H100 product call. We are on track to launch our first ever data center CPU, Grace, in the first half of twenty twenty three.

Speaker 2

Grace is the ideal CPU for AI factories. This week at Computex, we announced that dozens of server models based on Grace will be brought to market success by the first wave of system builders, including ASUS, Foxconn, Gigabyte, QCT, Supermicro and WiWin. These servers will be powered by the NVIDIA Grace CPU Superchip, which features 2 CPUs and the Grace Upper Superchip, which pairs an NVIDIA upper GPU with an NVIDIA Grace CPU in an integrated model. We've introduced new reference designs based on grace OVX for Digital Twins or Omniverse and HDX for HPC and AI. These server designs are all optimized for NVIDIA's rich accelerated computing software stacks and can be qualified as part of our NVIDIA certified systems lineup.

Speaker 2

The enabler for the Grace Hopper and Grace Superchips is our ultra energy efficient, low latency, high speed memory coherent interconnect called NVLink, which scales from die to die, chip to chip and system to system. Within the link, we can configure Grace and Hopper to address a broad range of workloads. Future NVIDIA chips, The CPUs, GPUs, DPUs, NICs and SoCs will integrate NVLink just like Grace and Hopper capabilities based on our world class SerDes technology. We are making NVLink open to customers and partners to implement custom chips platform. In networking, we're kicking off a major product cycle with the introduction of Spectrum 4, Spectrum 4 Switch, ConnectX7 SmartNIC, BlueField 3 DPU and the DOCA software.

Speaker 2

Built for AI, NVIDIA Spectrum 4 arrives as data centers are growing exponentially and demanding extreme performance, advanced security and powerful features to enable high performance, advanced virtualization and simulation at scale. Progress. Across our businesses, we are launching multiple new GPU, CPU, DPU and SoC products over the coming quarters. Moving to the rest of the P and L. GAAP gross margin for the Q1 was 65.5% and non GAAP gross margin was up 67.1%, up 90 basis points from a year ago and up 10 basis points sequentially.

Speaker 2

We have been able to offset rising costs and supply chain pressures. We expect to maintain gross margins at current levels in Q2. Q1. Going forward, as new products ramp and software becomes a larger percent of revenue, we have opportunities to increase gross margins longer term. GAAP operating margin was 22.5% impacted by a 1 point $35,000,000,000 acquisition termination charge related to the ARM transaction.

Speaker 2

Non GAAP operating margin was 47.7 sales and marketing expense. We are closely managing our operating expenses to balance the current macro environment with our growth opportunities and we've been very successful in hiring so far this year and are now slowing to integrate these new employees. This also enables us to focus our budget on taking care of our existing employees Q1 results. We are still on track to grow our non GAAP operating expenses in the high 20s range this year. We expect sequential increases to level off after Q2 as the first half of the year includes a significant amount of expenses related to the bring up Q1 results, which should not reoccur in the second half.

Speaker 2

During Q1, we repurchased $2,000,000,000 of our stock. Our Board of Directors increased and extended our share repurchase program to repurchase an additional common stock quarterly dividend payment plan. We are now ready to take our next question from the line of Q2 of fiscal 2023. Our outlook assumes an estimated impact of approximately 500,000,000 Q2, and we estimate the impact of lower sell through in Russia and China Q2 gaming sell in by $400,000,000 Furthermore, we estimate the absence of sales to Russia to have $100,000,000 impact on Q2 in data center. We expect strong sequential growth in data center and automotive Q1 results to be more than an offset by the sequential decline in gaming.

Speaker 2

Revenue is expected to be $8,100,000,000 plus or minus 2%. GAAP and non GAAP gross margins are expected to be 65.1% and 67.1 percent, respectively, plus or minus 50 basis points. GAAP operating expenses are expected to be $2,460,000,000 non GAAP operating expenses are expected to be 1,750,000,000 GAAP and non GAAP other income and expenses are expected to be an expense of approximately $40,000,000 excluding gains and losses nonaffiliated investments. GAAP and non GAAP tax rates are expected to be 12.5% plus or minus 1% excluding discrete items and capital expenditures are expected to be approximately $400,000,000 to 450,000,000 financial details are included in the CFO commentary and other information available on our IR website. Conference call.

Speaker 2

In closing, let me highlight the upcoming events for the financial community. We'll be attending the BofA Securities technology conference in person on June 7, where Jensen will participate in a keynote fireside chat. Earnings call to discuss the results of our Q2 of fiscal 2023 is scheduled for Wednesday, August 24. We will now open the call for questions. Operator, could you please poll for questions?

Speaker 2

Thank you.

Speaker 4

Thank you.

Operator

We ask that you please limit yourself to one question. We'll pause for just a moment to compile the Q and A roster. We'll take our first question from C. J. Muse with Evercore ISI.

Operator

Your line is open.

Speaker 5

J. Muse:] Yes,

Speaker 6

good afternoon. Thank you for taking the question. I guess, we'd love to get an update on how you're thinking about the gaming cycle from here. The business has essentially doubled over the last 2 years, And now we've got some crosswinds with crypto falling off, channel potentially clearing ahead of a new product cycle. You talked about macro challenges.

Speaker 6

But at the same time, only a third of the installed base, has RTX, and we're moving out from under supply. So we'd love to hear your thoughts From here, once we get beyond kind of the challenges around COVID lockdown in the July quarter, how are you thinking about gaming trends?

Speaker 3

Yes, CJ, thanks for the question. You captured a lot of the dynamics well in your question. The underlying dynamics of the gaming industry is really solid. Net of the situation with COVID lockdown in China and Russia. The rest of the market is fairly robust and We expect the gaming dynamics to be intact.

Speaker 3

There are several things that are driving the gaming industry. In the last 2 years alone, 100,000,000 new gamers came into the PC industry. The format has expanded tremendously to be an influencer as a platform for themselves, use it for broadcast. So many people are now using their home PCs as their second workstation, if you will, second studio, Because they're also working from home. It is our primary way of communicating these days.

Speaker 3

The need for GeForce PCs have never been greater. And so I think that the fundamental dynamics are really good. And so as we look into Q2. It's hard to predict exactly what when COVID and the Warren Rushes going to

Operator

Next, we'll go to Matt Ramsay with Cowen. Your line is open.

Speaker 7

Thank you very much. Good afternoon. Jensen, I wanted to ask a bit of a question on the data center business. In this upcoming cycle with H100. There's some IO upgrades that are happening in servers that I think are going to be a fairly strong driver for you in addition to what's going on With Hopper and the huge performance fleets that are there.

Speaker 7

I wanted to ask a longer term question though around your move to NVLink With Grace and Hopper and what's going on with your whole portfolio. Do you envision the business continuing to be sort of card driven attached To 3rd party servers or do you think revenue shifts dramatically or in a small way over time to be more sort of vertically integrated All of the chips together on NVLink and how is the industry sort of responding to that potential move? Thanks.

Speaker 3

Yes. I appreciate the question. The, let's see. The first point that you made is a very big point. The next generation service that are being teed up right now are all Gen 5.

Speaker 3

The IO performance is substantially higher than what was available before. And so you're going to see a pretty large refresh as a result of that. Brand new networking cards from our company and others. Gen 5, And so we're perfectly timed to ramp into the Gen 5 generation with Hopper. There are a lot of different system configurations you want to make.

Speaker 3

If you take a step back and look at

Speaker 4

the type of systems

Speaker 3

Done in the cloud for hyperscale nature, done on prem for enterprise computing, done at the edge. Each one of these workloads and deployment locations. The way that you manage would dictate a different system architecture. So there isn't one size that fits all, which is one of the reasons why It's so terrific that we support PCI Express that we innovated chip to chip interconnect for the very Before anybody else did, this is Nelson 7 years ago. We're in our 4th generation of NVLink that allows us to connect 2 chips next to each other, 2 dies, 2 chips, 2 modules, 2 SXM modules to 2 systems, 2 multiple systems.

Speaker 3

And so our coherent chip to chip link, NVLink, has made it possible for us to mix and match chips, dies, packages, systems and all of these different types of configurations. And I think that over time, You're going to see even more types of configurations. And the reason for that has to do with a couple of very important new type of data centers that are emerging And you start you're starting to see that now with fairly large installations, infrastructures with NVIDIA HPC and NVIDIA AI. These are really AI factories where you're processing the data, refining the data and turning that data into intelligence. These AI factories are essentially running 1 major workload and they're running it 20 fourseven.

Speaker 3

Deep Recommender Systems is a good example of that. In the future, you're going to see large language models essentially becoming a platform themselves that would be running 20 fourseven, hosting a whole bunch of applications. And then on the other end, you're seeing data centers at the edge that are going to be robotics or autonomous data centers that are running 20 fourseven. They are going to be running in factories and retail stores and warehouses, logistic warehouses, all over the world. So these two new type of data centers Are just emerging and they also have different architectures.

Speaker 3

So, I think the net of it all is that our ability to support every single workload because we have a universal accelerator running every single workload from data processing data analytics to high performance computing to training to inference that we can support ARM and x86, that capability for us is makes it possible for us to really be able to serve all of these different segments. With respect to vertical integration, I think that system integration, The better way of maybe saying that is that system integration is going to come in all kinds of different ways. We're going to do semi custom chips as we've done with many companies in the past, including Nintendo. We'll do semi custom chiplets as we do with EnbLink. EnbLink is open to our partners and they could bring it to any fab and connect it coherently into our chip.

Speaker 3

We could do multi module packages. We could do multi package systems. So there's a lot of different ways to do system integration.

Operator

Next, we'll go to Stacy Rasgon with Bernstein Research. Your line is now open.

Speaker 8

Hi, guys. Thanks for taking my question. I wanted to follow-up on the sequential. So, Colette, I know you said the $500,000,000 was a $400,000,000 hit to gaming and a $100,000,000 hit to data. So I'm assuming that That doesn't mean the gaming is down $400,000,000 I mean is gaming do you see gaming actually down more than the actual Russia and lockdown hit?

Speaker 8

And I guess just how do I think about the relative sequentials of the businesses in light of those constraints that you guys are facing?

Speaker 2

Sure. Let me start first with, what does that mean to gaming? What does that mean to gaming for Q2? We do expect gaming to decline into Q2. We still believe our end demand remains very strong.

Speaker 2

Ampere It's just been a great architecture and there's many areas where we continue to see strength and growth in both our sell through and probably what we will see added into that channel as well. But in total, Q2 gaming will decline from last quarter from Q1 that it will probably decline in the teens, as we try and work through some of these lockdowns in China, which are holding us up. So overall, the demand for gaming is still strong. We still expect end demand to grow year over year in Q2.

Operator

Presentation. Next, we'll go to Mark Lipacis with Jefferies. Your line is open.

Speaker 9

Hi. Thanks for taking my question. If you listen to the networking OEMs this earnings season, it seems that there was a lot of talk about increased spending enterprise on their data centers and sometimes you hear them talking about how this is being driven by AI. You talked about your year over year growth in your cloud versus enterprise. Spending.

Speaker 9

I wonder if you could talk about what you were seeing sequentially. Are you seeing a sequential inflection in the enterprise? And can you talk about the attach rate of software for enterprise versus data centers and what which software is are you seeing the most interest? I know you talked about is it Omniverse? Is it Natural Language Processing?

Speaker 9

Or Is there one big driver or is there a bunch of drivers for the various different software packages you have? Thank you.

Speaker 3

Yes. Thanks, Mark. We had a record data center business this last quarter. We expect to have a record another record quarter this quarter Square and extracting insight from the vast amount of data that companies have It's incredibly strategic to all the companies that we know, because in the final analysis, AI is about Automation of Intelligence and most companies are about domain specific intelligence. We want to produce intelligence.

Speaker 3

And there are several techniques now that have been created to make it possible for most companies to apply their data to extract insight and to automate a lot of the predictive things that they have to do and do it quickly. And so I think the trend that you hear other people experiencing about machine learning, data analytics, data driven insights, artificial intelligence, however it's described, it's all exactly the same thing. And it's sweeping just about every industry and every company. Our networking business is also highly supply constrained. Our demand is really, really high.

Speaker 3

And it requires a lot of components aside from just our chips. Components and transceivers and connectors and cables and just it's a really it's a complicated system, network and there are many physical components. And so the supply chain has been problematic. We're doing our best and Our supply has been increasing from Q4 to Q1. We're expecting it to increase in Q2 and increase in Q3 and Q4.

Speaker 3

And so we're really grateful for the support from the component industry around us And we'll be able to increase that. With respect to software, there are 2 first of all, there are all kinds of machine learning models, computer vision, speech AI, natural language understanding, all kinds of robotics applications, the most probably the largest, the most visible one is self driving cars, which is essentially a robotic AI. And then recently, this incredible breakthrough from AI model called transformers That has led to really, really significant advances in natural language understanding. And so, there are all these different types of models. There are thousands and thousands of species of AI models and in use in all these different industries.

Speaker 3

One of my favorite, I'll just say very quickly and I'll answer that question about the software. One of my favorites is using transformers to understand the language of chemistry. We're using transformers and using AI models to understand the language of proteins, amino acids, which is genomics. To apply AI to understand to recognize the patterns, to understand the sequence And essentially understand the language of chemistry and biology is a really, really important breakthrough. And all of this excitement Around synthetic biology, much of it stands back to some of these inventions.

Speaker 3

But anyhow, all of these different models need an engine to run on. And that engine is called NVIDIA AI. In the case of hyperscalers, They can cobble together a lot of open source and we provide a lot of our source to them and a lot of our engines to them for them to operate their AI. But for enterprises, they need someone to package it together and be able to support it and refresh it, update it for new architectures, support old architectures and their installed base, etcetera, and all the different use cases that they have. And so that engine is called NVIDIA AI.

Speaker 3

It's almost like a sequel engine if you will, and except this is an engine for artificial intelligence. There's another engine that we provide And that engine is called Omniverse and it's designed for the next wave of AI, where artificial intelligence has to not just manipulate information like recommender systems and conversational systems and such, But it has to interact with physical systems, whether it's interacting with physics directly, meaning robotics or being able to automate physical systems like heat recovery steam generators, which is really important today. And so, Omniverse is designed to be able to sit at that interface, the intersection between Simulation and Artificial Intelligence and that's what Omniverse is about. Omniverse has now, let's see, some we're still early in deployment of Omniverse for commercial license. It's been a couple of quarters now since we've released Omniverse Enterprise.

Speaker 3

And I think at this point we have 10% of the world's top 100 companies that are already customers, licensing customers. Substantially more who are evaluating. I think it's been downloaded nearly 200,000 times. It is being tried in some 700 companies and Colette highlighted some of the companies. You might some of the companies that are using it in all kinds of interesting applications at GTC.

Speaker 3

And so I fully expect that The NVIDIA AI engine, the Omniverse engine are going to be very successful for us in the future and contribute greatly to our earnings.

Operator

Next, we'll go to Vivek Arya with BofA Securities. Your line is open.

Speaker 10

Thanks. Just wanted to clarify, Colette, if your Q2 outlook includes any destocking benefits from the new products That you are planning to launch this year. And then, Jensen, my question is for you. You're still guiding data center to a very strong, I think close to 70% or so year on year growth despite all the headwinds. Are you worried at all about all the headlines about the slowdown in the macro economy?

Speaker 10

Vivek. Is there any cyclical impact on data center growth that we should keep in mind as we think about the second half of the year?

Speaker 2

Yes. Vivek, let me first answer the question that you asked regarding any new products Q2. As we discuss about it, most of the ramp that we have of our new architectures, we're going to see in the back half of the year. We're going to start to see, for example, Hopper will probably be here in Q3, but starting to ramp closer to the end of the calendar year. So you should think about most of our product launches to be ramping in the second half of the year on that part.

Speaker 2

I'll turn it over for Jenson for the rest.

Speaker 3

Thanks. Our data center Demand is strong and remains strong. Hyperscale and cloud computing revenues, as you mentioned, has grown significantly, has doubled year over year. And we're seeing really strong adoption of A100. A100 is really quite special and unique in the world of accelerators.

Speaker 3

And this is one of the really, really great innovations as we extended our GPU from graphics to CUDA to Tensor Core GPUs. It's now a universal accelerator and so you could use it for data processing for ETL for example extract, transform and load. You could use it for database acceleration. Many SQL functions are accelerated on NVIDIA GPUs. We accelerate RAPIDS, we accelerate, which is the Python version, data center scale version of Pandas.

Speaker 3

We accelerate Spark 3.0. And so from database queries to data processing to extraction and transform and loading of data before you do training and inference and whatever image processing or other algorithmic can be fully accelerated on a100. And so we're seeing great success there. At the core and closer Q1. To what is happening today, you're seeing several different very important new AI models that are being invested in at very, very large scale and with great urgency.

Speaker 3

You probably have heard about Deep Recommender Systems. This is the economic engine, the information filtering engine of the Internet. If not for the recommender system, it would be practically impossible for us to enjoy our Internet experience, shopping experience with trillions of things that are changing in the world every day incredible thing called a recommender system. The second thing is conversational AI. You're seeing chatbots and website customer service, even live customer service being now supported by AI, conversational AI, has an opportunity to enhance the customer service on the one hand, on the other hand, supplement for a lot of labor shortage.

Speaker 3

And then the third is this groundbreaking piece of work as related to transformers that led to natural language understanding breakthrough, but within it is this incredible thing called large language models, which embeds human knowledge because it's been trained in so much data and we recently announced MEGATRON 530B And it was a collaboration we did with Microsoft, the foundation of I think they call it Turing. And this language model and others like it, like OpenAI's GPD3, are really transformative and they take an enormous amount of computation. However, the net result is a pretrained model that is really quite remarkable. Now we're working with thousands of startups, large companies that are building, who are using the public cloud and so it's driving a lot of demand for us in the public cloud. I think we have now 10,000 AI inception startups that are working with us and using NVIDIA AI, a lot better and they could do greater things.

Speaker 3

And so that's driving AI in the cloud. And so all of these different factors, whether it's just The industrial recognition of the importance of AI, the transformative nature of these new AI models, recommender systems, large language models, conversational AI, the thousands of companies around the world that are using NVIDIA AI in the cloud, driving public cloud demand. All of these things are driving our data center growth. And so we expect to see data center demand remain strong.

Operator

Next we'll go to Tim Arcuri with UBS. Your line is open.

Speaker 11

Call. Thank you very much. I had a question about this $500,000,000 impact for July and whether it's more supply related or demand related. And that's because most others in Sunnys are sort of citing this China stuff in particular as more of a logistics issue, so more of a supply issue. But the language Colette you were using in your commentary cited lower sell through in gaming and sort of the absence of sales in Russia.

Speaker 11

To me that sounds a little more demand, Which would make sense in the context of this new, freeze on hiring that you have. So I ask because if it's supply related, then you could argue that it's not perishable and really just timing. But if demand related, that might never come back and It could be the beginning of a falling knife. So, I wonder if you can sort of walk through that for me. Thanks.

Speaker 2

Thanks, Tim, for the question. Let me try and bet here on the China and Russia, 2 very different things. The current China lockdowns challenges in terms of the logistics throughout the country, things going in, out of the country. It puts a lot of pressure logistics that were already under pressure. From a demand perspective, it has also been hit from the gaming side.

Speaker 2

You have very large cities that are in full lockdown, focusing really on other important things for the citizens there. So it's impacting our demand. We do believe that they will come out The supply will sort it out. It's very difficult to determine how. Now in the case of Russia, we're not selling to Russia.

Speaker 2

That's something that we had announced earlier last quarter. But there were plans and Russia has been a part of our overall company revenue, probably about 2% of our company revenue historically and a little larger percentage when you look at our gaming business. Hope that

Speaker 4

helped.

Operator

Next, we'll go to Ambrish Srivastava with BMO. Your line is now open.

Speaker 4

Hi, thank you very much, Colette and gentlemen. I actually really appreciate it that you call out demand for most chip companies. It seems like it's heresy to say Demand is a problem, so refreshing to hear that. I had a question on the second half, and it relates to both data center as well as gaming. So last couple of times you have talked publicly, you have made comments that your visibility into data center has never been better.

Speaker 4

So I was wondering if you just take out the Russia impact. Is that still true, all the orders that you have been getting their intact and you did say that business will see a strong momentum. I just wanted to make sure that statement of confidence you have made stays. And then on gaming, Colette, do we expect second half to be up year over year Q2. It seems like it could be up sequentially, but may not return to year over year growth in Q3.

Speaker 4

Thank you.

Speaker 3

Yes. Ambrish, thanks for the question. On first principles, vastly better than a couple of years ago. And the reason for that is several. One, if you recall a couple of 2, 3 years ago, Deep Learning and AI was starting to accelerate in the most computer science deep companies in the world with CSPs and hyperscalers.

Speaker 3

And but just about everywhere else, it was still quite nascent. And there was a couple of reasons for that. Obviously, the understanding of the technology is not as pervasive at the time. The type of industrial use cases for artificial intelligence requires labeling of data that's really quite difficult. And then now, with transformers, you have unsupervised learning and other techniques, Zero Shot Learning that allows us to do all kinds of interesting things without having to have human labeled data.

Speaker 3

We even have synthetic generated data with Omniverse that helps customers do data generation without having to label data, which is either too cost effect, too costly Quite frankly, oftentimes impossible. And so now the knowledge and fairly effective way and in many industries rather transformative. And so I think number 1, We went from clouds as hyperscalers to all of industries. 2nd, we went from trading focused to inference. Most people thought that inference was going to be easy.

Speaker 3

It turns out inference is by far harder. And the reason for that is because there's so many different models There's so many different use cases and so many quality of service requirements and you want to run these inference models in a small of a footprint as you can. And so, When you scale out, the number of users that use the service is really quite high. So, using acceleration and using NVIDIA's platform. We could inference any model from computer vision to speech to chemistry to biology, you name it.

Speaker 3

And we do it so quickly and so fast that the cost is very low. And so the more acceleration you do, The more money you will save and that I think that wisdom is absolutely true. And so the second dimension is training to inference. The 3rd dimension is that we now have so many different types of configurations of systems that we can go from high performance computing systems all the way to cloud, to on prem, to edge. And then the final concept is really this industrial deployment now of AI that's causing us to be able to in just about every industry find growth.

Speaker 3

And so, as you know, our cloud and hyperscalers are growing very, very quickly. However, the vertical part vertical industries, which is the financial services, retail and telco and all of those vertical industries have also grown very, very nicely. And so in all of those different dimensions, our visibility should be a lot better. And then starting a couple of years ago, adding the Mellanox portfolio to our company, we're able to provide a lot more solution oriented end to end platform solutions And so our networking business is growing very, very nicely as well.

Operator

Next, we'll go to Harlan Sur with JPMorgan. Your line is open.

Speaker 12

Hi, good afternoon. Thanks for letting me ask a question. I just want to maybe just ask this question a little bit more directly. So it's good to see the team being able to drive navigate the dynamic supply chain environment, right? You strong sequential growth in data center in April here in the July quarter, even with some demand impact from Russia, right?

Speaker 12

And so as we think about the second half of Cloud spending is strong and it's actually, I think accelerating. You're getting ready to ramp H100 later in the year. Novanox, I think, is getting more supply as you move through the year. And in general, I think previously, you guys were anticipating sequential supply and revenue growth

Speaker 3

Either one of us can The answer is yes. We see a strong demand in data center, hyperscale to cloud computing to vertical industries. Ampere is going to continue to scale out. It's been qualified in every single company in the world. And so after 2 years, it remains the best universal accelerator on the planet and it's going to continue to scale out in all these different domains and different markets.

Speaker 3

We're going to layer on top of that a brand new architecture Hopper. We're going to layer on top of that brand new networking architectures, Quantum 3, CX7, BlueField 3 and we have increasing supply. And so we're looking forward to an excellent quarter, next quarter again for data centers

Operator

conference call. Next, we'll go to Chris Caso with Raymond James. Your line is open.

Speaker 13

Yes. Thank you. Wonder if you could speak a little bit about the purchase obligations, which seem like they were up again in the quarter. And how that Was that a function of longer dated obligations or a higher magnitude of obligations? And maybe you could just speak to supply constraints in general.

Speaker 13

You've mentioned a couple of times in the call about continued constraints in networking business. What about the other parts of business. Where are you still constrained?

Speaker 2

Yes. So let me start here and I'll see if Jensen wants to add more of that. Our purchase obligations as well as our prepaids have 2 major things to keep in mind. 1, for the first time ever, We are prepaying to make sure that we have that supply and those commitments long term. And additionally, on our purchase obligations, many of them are for long lead time items that are a must for us to procure to make sure that we have the products coming to market.

Speaker 2

A good percentage of our purchase commitments is for our data center business, which you can imagine are much larger systems, much more complex systems and those things that we are procuring to make sure we can feed the demand Q1 results both in the upcoming quarters and further. Areas in terms of where, we are still a little bit supply constrained. Our networking, our demand is quite strong. We've been improving it each time, but yes, we still have demand excuse me, supply concerns With networking still. Is there others that you want to add on, Jensen?

Speaker 3

No, I thought you were perfect. That's

Operator

perfect. Our final question comes from Aaron Rakers with Wells Fargo. Your line is open.

Speaker 5

Yes. Thanks for fitting me in. And most of my questions around gaming and beta finner have been answered. But I guess I'll ask about the auto segment. While it's still small, Clearly, you guys sound confident in that business starting to see significant sequential growth into this next quarter.

Speaker 5

I wonder if you could help us kind of think about the trajectory of that business over the next couple of quarters. And I think in the past, you've said that that Should start to really inflect higher as we move into the second half of the year. Just curious if you can help us think about that piece of the business?

Speaker 3

Several data points. We are just starting. We have just started shipping Orin and the Q1 of shipping production Orin. Orin is a robotics processor. It's designed for a software defined robotic card or robotic pick and placer or robotic mover, logistics mover.

Speaker 3

We've been designed into 35 car and trucks and robotaxi companies and more others if you include logistics movers and last mile delivery systems and farming equipment. And The number of design wins for Orin is really quite fantastic. Orin is a revolutionary processor and it's designed as a, if you will, data center on a chip. And it is the 1st data center on a chip that is robotic, processes sensor information. It's safe.

Speaker 3

It has the ability to be rather resilient. It has confidential computing. It is designed to be secure, designed to be all those things, because these data centers are going to be everywhere. And so, Oren is really a technological marvels production. We experienced very likely the lowest auto quarter in some time for some time.

Speaker 3

And the reason for that is because over the next 6 years or so, we have $11,000,000,000 and counting business that we've secured estimated. And so I think it's a fairly safe thing to say now that Orin and our autonomous vehicle and robotics business It's going to be our next multi $1,000,000,000 business. It's on its way surely there. The robotics and autonomous systems and autonomous machines, whether they move or not move, but AI systems that are at the physical edge. It's surely going to be the next major computing segment.

Speaker 3

It is surely going to be the next major data center segment. We've been working in this area as you know for a decade. We have a fair amount of expertise in this area And Oren is just one example of our work here. We have 4 pillars to our strategy for autonomous systems, Starting from the data processing and the AI training part of it, to train robotics AIs. Memory of the robotics AI otherwise known as mapping.

Speaker 3

And then finally, the actual robotics application and the robotics processor in the system and that's where Orin goes. But Orin is just one of our 4 pillars I'm really enthusiastic about the next phase of the computer industry's growth and I think a lot of it's going to be at the edge, a lot of it's going to be about robotics.

Operator

Call. Thank you. I'll now turn it back over to Jensen Huang for any additional closing remarks.

Speaker 3

Thanks, everyone. The full impact and duration of the war in Ukraine and COVID lockdowns in China difficult to predict. However, the impact of our technology and our market opportunities remain unchanged. Sequential and temporal complexity. Researchers are creating transformer models that are revolutionizing applications From Robotics to Drug Discovery.

Speaker 3

The effectiveness of deep learning AI is driving companies across industries to adopt NVIDIA for AI Computing. Conference call. We're focused on 4 major initiatives. 1st, ramping our next generation of AI infrastructure chips and platforms, Hopper GPU, BlueField DPU, NVLink, InfiniBand, Quantum InfiniBand Spectrum Ethernet Networking, And all this to help customers build their AI factories and take advantage of new AI breakthroughs like transformers. 2nd, ramping our system and software industry partners to launch Grace, our first CPU.

Speaker 3

3rd, ramping Orin, our new robotics processor and nearly 40 customers building cars, robo taxis, trucks, delivery robots, logistics robots, farming robots to medical instruments. And 4th, software platforms, adding new value to our ecosystem with NVIDIA AI and NVIDIA Omniverse and expanding into new markets with new CUDA acceleration libraries. These initiatives will greatly advance AI. And while continuing to extend this most impactful technology of our time scientific expertise and expertise in every field and companies in every industry. We look forward to updating you on our progress next quarter.

Operator

Conference call. This concludes today's conference call. You may now disconnect.