AMD wants people to remember that Nvidia’s not the only company selling AI chips. It’s announced the availability of new accelerators and processors geared toward running large language models, or LLMs.
AMD releases new chips to power faster AI training

The chipmaker unveiled the Instinct MI300X accelerator and the Instinct M1300A accelerated processing unit (APU), which the company said works to train and run LLMs. The company said the MI300X has 1.5 times more memory capacity than the previous M1250X version. Both new products have better memory capacity and are more energy-efficient than their predecessors, said AMD.
“LLMs continue to increase in size and complexity, requiring massive amounts of memory and compute,” AMD CEO Lisa Su said. “And we know the availability of GPUs is the single most important driver of AI adoption.”
Su said during a presentation that MI300X “is the highest performing accelerator in the world.” She claimed MI300X is comparable to Nvidia’s H100 chips in training LLMs but performs better on the inference side — 1.4 times better than H100 when working with Meta’s Llama 2, a 70 billion parameter LLM.
AMD partnered with Microsoft to put MI300X in its Azure virtual machines. Microsoft CTO Kevin Scott, a guest during Su’s speech, also announced the Azure ND MI300X virtual machines — first revealed in November — are now available on preview. Meta also announced it will deploy MI300 processors in its data centers.
Su said AMD released the MI300A APU for data centers, which she said are expected to grow its total addressable market to $45 billion. APUs generally combine CPUs and GPUs for faster processing. AMD said the MI300A offers higher-performance computing, faster model training, and a 30 times energy efficiency improvement. Compared to the H100, AMD said it has 1.6 times the memory capacity. It also features unified memory, so there is no need to move data from different devices anymore.
MI300A will power the El Capitan supercomputer built by Hewlett Packard Enterprise at the Lawrence Livermore National Laboratory. El Capitan is considered one of the most powerful supercomputers and is expected to deliver more than two exaflops of performance.
The MI300A APU “is now in production and is being built into data centers.”
Pricing information was not immediately available.
Su teased the MI300 chips during the Code Conference, saying AMD was excited about the opportunity to tap more chip users, not just from cloud providers but from enterprises and startups.
AMD also announced the latest addition to its Ryzen processors, the Ryzen 8040, which can put more native AI functions into mobile devices. The company said the 8040 series offers 1.6 times more AI processing performance than previous models and integrates neural processing units (NPUs).
The company said Ryzen 8040 would not be limited to AI processing, as it claimed video editing would be 65 percent faster and gaming would be 77 percent faster than with competing products like Intel’s chips.
AMD expects manufacturers like Acer, Asus, Dell, HP, Lenovo, and Razer to release products integrating Ryzen 8040 chips in the first quarter of 2024.
Su said the next generation of its Strix Point NPUs will be released in 2024.
AMD also announced the Ryzen AI Software Platform is now widely available, which will let developers building AI models on Ryzen-powered laptops offload models into the NPU so the CPU can reduce power consumption. Users will get support for foundation models like the speech recognition model Whisper and LLMs like Llama 2.
To power AI models — and take advantage of the current hype for the tech — companies like AMD, Nvidia, and Intel have been locked in what’s basically an AI chip arms race. So far, Nvidia captured the largest market share with its highly coveted H100 GPUs used to train models like OpenAI’s GPT.
AMD wants people to remember that Nvidia’s not the only company selling AI chips. It’s announced the availability of new accelerators and processors geared toward running large language models, or LLMs. The chipmaker unveiled the Instinct MI300X accelerator and the Instinct M1300A accelerated processing unit (APU), which the company said…
Recent Posts
Archives
- February 2025
- January 2025
- December 2024
- November 2024
- October 2024
- September 2024
- August 2024
- July 2024
- June 2024
- May 2024
- April 2024
- March 2024
- February 2024
- January 2024
- December 2023
- November 2023
- October 2023
- September 2023
- August 2023
- July 2023
- June 2023
- May 2023
- April 2023
- March 2023
- February 2023
- January 2023
- December 2022
- November 2022
- October 2022
- September 2022
- August 2022
- July 2022
- June 2022
- May 2022
- April 2022
- March 2022
- February 2022
- January 2022
- December 2021
- November 2021
- October 2021
- September 2021
- August 2021
- July 2021
- June 2021
- May 2021
- April 2021
- March 2021
- February 2021
- January 2021
- December 2020
- November 2020
- October 2020
- September 2020
- August 2020
- July 2020
- June 2020
- May 2020
- April 2020
- March 2020
- February 2020
- January 2020
- December 2019
- November 2019
- September 2018
- October 2017
- December 2011
- August 2010