AMD gears up for an epic showdown against Nvidia in AI chips
- AMD sees strong demand for its new Instinct MI300 GPUs.
- AMD claims are the amd Instinct MI300 chips are the highest-performance accelerators in the world for generative AI.
- Lisa Su foresees the AI chip market surpassing $400 billion by 2027.
In the dynamic graphics processing units (GPUs) sector, AMD and Nvidia have long stood as the leading contenders, engaging in fierce and enduring competition. With decades of rivalry behind them, the intensity of their contest has only heightened over the years. This year, AMD elevated its competitive stance by acquiring Nod.AI as part of its “AI growth strategy.” The rivalry reached a new level this week as AMD unveiled its latest line of AI chips – the Instinct MI300 – directly challenging Nvidia’s supremacy in the field of AI.
At a mid-week “Advancing AI” event in San Jose, California, AMD said it will go head to head with Nvidia in the AI chips market by launching a new lineup of accelerator processors. The AMD Instinct™ MI300 Series accelerators, according to CEO Lisa Su, are designed to outperform rival products in running AI software.
The launch marks a pivotal moment in AMD’s five-decade history, positioning the company for a significant face-off with Nvidia in the thriving AI accelerators market. These chips play a crucial role in AI model development, excelling at processing data compared to conventional computer processors.
“The truth is we’re so early,” Su said, according to Bloomberg‘s report. “This is not a fad. I believe it.” The event signals that AMD is growing confident that its MI300 lineup can attract major tech players, potentially redirecting significant investments to the company. During the Wednesday AMD investor event, Meta, OpenAI, and Microsoft announced their adoption of AMD’s latest AI chip, the Instinct MI300X.
The move signals a notable shift among tech companies seeking alternatives to Nvidia’s costly graphics processors, which have traditionally been crucial for developing and deploying AI programs like OpenAI’s ChatGPT. On the same day, Nvidia experienced a 2.3% decline in shares to US$455.03 in New York, reflecting investor concerns about the perceived threat posed by the new chip.
However, AMD did not witness a proportional share surge, with a modest 1.3% decrease to $116.82 on a day when tech stocks were generally underperforming.
Nvidia’s market value surpassed US$1.1 trillion this year, fueled mainly by the increasing demand for its chips from data center operators.
However, the looming question around Nvidia’s dominance in the accelerator market is how long it can maintain its current position without significant competition. The launch of the AMD Instinct MI300 will likely test that question to its limits.
AMD vs Nvidia: who has better AI chips right now?
The MI300 series of processors, which AMD calls “accelerators,” was first announced nearly six months ago when Su detailed the chipmaker’s strategy for AI computing in the data center. AMD has built momentum since then.
This week’s announcement means that theMI300X and MI300A are shipping and in production. The MI300X, designed for cloud providers and enterprises, is built for generative AI applications and outperforms Nvidia’s H100 GPU in two key metrics: memory capacity and memory bandwidth, which lets AMD to deliver comparable AI training performance and significantly higher AI inferencing performance.
In other words, the MI300X GPU boasts over 150 billion transistors, surpassing Nvidia’s H100, the current market leader, with 2.4 times the memory. Additionally, it reaches 5.3 TB/s peak memory bandwidth, which is 1.6 times more than Nvidia’s H100’s 3.3 TB/s.
“AMD sees an opening: large language models — used by AI chatbots such as OpenAI’s ChatGPT — need a huge amount of computer memory, and that’s where the chipmaker believes it has an advantage,” a Bloomberg report reads.
“It’s the highest performance accelerator in the world for generative AI,” Su said at the company’s “Advancing AI” event. In a media briefing before the event, Brad McCredie, AMD’s corporate vice president of data center GPU, said the MI300X GPU will still have more memory bandwidth and memory capacity than Nvidia’s next-generation H200 GPU, which is expected next year.
In November, when Nvidia revealed its intentions for the H200, the company specified that the H200 would incorporate 141 GB of HBM3 memory and boast a memory bandwidth of 4.8 TB/s. Naturally then, McCredie expects fierce competition between the two companies on the GPU front.
“We do not expect (Nvidia) to stand still. We know how to go fast, too,” he said. “Nvidia has accelerated its roadmap once recently. We fully expect t to keep pushing hard. We will keep going hard, so we fully wish to stay ahead. Nvidia may leapfrog. We’ll leapfrog back.”
The AMD Instinct MI300A APU comes equipped with 128GB of HBM3 memory. Compared to the prior M250X processor, the MI300A delivers 1.9 times more performance per watt on HPC and AI workloads, according to the company.
Forrest Norrod, executive vice president and general manager of AMD’s data center solutions business group explained that at the node level, the MI300A achieves twice the HPC performance-per-watt compared to its closest competitor. That lets customers accommodate more nodes within their overall facility power budget, aligning with sustainability goals.
AMD’s outlook
In her address, Su emphasized that AI is the most transformative technology in the last 50 years, surpassing even the introduction of the internet. Notably, she highlighted the accelerated adoption rate of AI compared to other groundbreaking technologies. Last year, AMD projected a 50% growth in the data center AI accelerated market from US$30 billion in 2023 to US$150 billion in 2027.
Amazing day advancing #AI with our launch of @AMD Instinct MI300X – the most advanced AI solution in the industry. Huge thanks to our partners who joined us today… together we are building the future of AI. @Microsoft @OracleCloud @Meta @Dell @Supermicro_SMCI @HPE @Lenovo… pic.twitter.com/8TvCbQwCEl
— Lisa Su (@LisaSu) December 7, 2023
Su revised this estimate, now predicting an annual growth of more than 70% over the next four years, with the market reaching over US$400 billion in 2027. Su also underscored the pivotal role of generative AI, stressing the substantial investment required for new infrastructure to facilitate training and inference, describing the market as tremendously significant.