: Nvidia’s stock slips on introduction of new AI data-center CPU+GPU chip

United States

You didn’t think Nvidia Corp. would let the lead AI data-center product of Advanced Micro Devices Inc., its closest rival in the artificial-intelligence data-center chip market, go unanswered, did you?

At 2023’s Special Interest Group on Computer Graphics and Interactive Techniques, or SIGGRAPH, conference, Nvidia NVDA, -1.66% founder and Chief Executive Jensen Huang unveiled his answer to AMD’s AMD, -3.06% Insight MI300X CPU + GPU, which that company has been teasing all year.

In his keynote address Tuesday, Huang introduced the next-generation DGX GH200 Grace Hopper Superchip, for use with large-memory generative-AI models like OpenAI’s ChatGPT — which is backed by Microsoft Corp. MSFT, -1.23% — “to scale out the world’s data centers.”

Nvidia shares, which had been down about 1% before the announcement, dropped as much as 3% to an intraday low of $ 440.56 following the announcement, and finished down 1.7% at $ 446.64.

AMD shares closed down 3.1% at $ 113.23. Meanwhile, the PHLX Semiconductor Index SOX fell 1.6%, the S&P 500 SPX declined 0.4% and the tech-heavy Nasdaq Composite COMP dropped 0.8%.

In a press conference ahead of the announcement, Nvidia’s head of hyperscale and high-performance computing, Ian Buck, told reporters the GH200 packs more memory and more bandwidth than the company’s H100-based data-center system. The GH200 uses Nvidia’s Hopper GPU and marries it with its Arm Ltd. architecture-based Grace CPU. The chip carries 141 GB of HBM3 memory and 5 TB per second of bandwidth.

The GH200 can be doubled up in the NVLink-dual GH200 system to increase capacity by 3.5 times and triple bandwidth. Both will be available in the second quarter of 2024, but Nvidia did not comment on pricing.

Nvidia

Buck said a vast majority of AI training and inferencing is done on Nvidia’s current HGX systems. The GH200 offers a new option for inference customers to support AI workloads, at two times the performance per watt, as cloud-service providers hope to build out their capacity without significantly increasing their energy costs.

In addition to Microsoft’s Azure cloud-service provider, other companies with hyperscaler capacity, like Amazon.com Inc.’s AMZN, -1.60% giant Amazon Web Services and Alphabet Inc.’s GOOG, -0.08% GOOGL, -0.10% Google Cloud Platform, are expected to fuel sales of AI chips in buildouts in the second half of the year.

Read: Chip-equipment suppliers rally after Lam says AI servers will drive growth

On Friday, AMD broke with the broad tech selloff to finish higher on the week as analyst support gathered for the chip maker’s AI position following its earnings report, in which AMD Chair and CEO Lisa Su forecast “multiple winners” in the AI race. Nvidia reports its earnings after the market close on Aug. 23.

Read: Will AI do to Nvidia what the dot-com boom did to Sun Microsystems? Analysts compare current hype to past ones.

AMD introduced its MI300X CPU + GPU at its AI product launch in June. The chip maker is regarded as a distant second to Nvidia when it comes to AI data-center hardware market share.

Read: Nvidia gets more good news from Big Tech, even as AI spending ‘may not lift all boats’

Year to date, AMD shares have gained 74.8%, while Nvidia shares have soared more than 205% and the SOX index has rallied 45.3%. The S&P 500 has advanced 17.2% and the Nasdaq has grown 327% in the same time frame.

Read: Nvidia ‘should have at least 90%’ of AI chip market with AMD on its heels