(Bloomberg) — Advanced Micro Devices Inc., taking aim at a burgeoning market dominated by Nvidia Corp., unveiled new so-called accelerator chips that it said will be able to run artificial intelligence software faster than rival products.
Most Read from Bloomberg
The company introduced a long-anticipated lineup called the MI300 at an event Wednesday held in San Jose, California. Chief Executive Officer Lisa Su also gave an eye-popping forecast for the size of the AI chip industry, saying it could climb to more than $400 billion in the next four years. That’s more than twice as high as a projection AMD gave in August, showing how rapidly expectations are changing for AI hardware.
The launch is one of the most important in AMD’s five-decade history, setting up a showdown with Nvidia in the red-hot market for AI accelerators. Such chips help develop AI models by bombarding them with data, a task they handle more adeptly than traditional computer processors.
AMD is showing increasing confidence that the MI300 lineup can win over some of the biggest names in technology, potentially diverting billions in spending toward the company. Customers using the processors will include Microsoft Corp., Oracle Corp. and Meta Platforms Inc., AMD said.
Nvidia shares dipped 1.4% to $458.98 in New York on Wednesday, while AMD’s stock fell less than 1% to $117.24.
Surging demand for Nvidia chips by data center operators helped propel its shares this year, sending the company’s market value past $1.1 trillion. The big question is how long it will essentially have the accelerator market to itself.
AMD sees an opening: Large language models — used by AI chatbots such as OpenAI’s ChatGPT — need a huge amount of computer memory, and that’s where the chipmaker believes it has an advantage.
The new AMD chip has more than 150 billion transistors and 2.4 times as much memory as Nvidia’s H100, the current market leader. It also has 1.6 as much memory bandwidth, further boosting performance, AMD said.
Su said that the new chip is equal to Nvidia’s H100 in its ability to train AI software and much better at inference — the process of running that software once it’s ready for real-world use.
But Nvidia is developing its own next-generation chips. The H100 will be succeeded by the H200 in the first half of next year, giving access to a new high-speed type of memory. That should match at least some of what AMD’s offering. And then Nvidia is expected to come out with a whole new architecture for the processor later in the year.
AMD’s prediction that AI processors will grow into a $400 billion market underscores the boundless optimism in the artificial intelligence industry. That compares with $597 billion for the entire chip industry in 2022, according to IDC.
As recently as August, AMD had offered a more modest forecast of $150 billion over the same period. But it will take the company a while to grab a large piece of that market. AMD has said that its own revenue from accelerators will top $2 billion in 2024, with analysts estimating that the chipmaker’s total sales will reach about $26.5 billion.
The chips are based on the type of semiconductors called graphics processing units, or GPUs, which have typically been used by video gamers to get the most realistic experience. Their ability to perform a certain type of calculation rapidly by doing many of computations simultaneously has made them the go-to choice for training AI software.
(Updates shares in fifth paragraph.)
Most Read from Bloomberg Businessweek
©2023 Bloomberg L.P.