AMD desires folks to do not forget that Nvidia’s not the one firm promoting AI chips. It’s introduced the provision of recent accelerators and processors geared towards working giant language fashions, or LLMs.
The chipmaker unveiled the Intuition MI300X accelerator and the Intuition M1300A accelerated processing unit (APU), which the corporate stated works to coach and run LLMs. The corporate stated the MI300X has 1.5 instances extra reminiscence capability than the earlier M1250X model. Each new merchandise have higher reminiscence capability and are extra energy-efficient than their predecessors, stated AMD.
“LLMs proceed to extend in measurement and complexity, requiring huge quantities of reminiscence and compute,” AMD CEO Lisa Su stated. “And we all know the provision of GPUs is the only most necessary driver of AI adoption.”
Su stated throughout a presentation that MI300X “is the very best performing accelerator on this planet.” She claimed MI300X is corresponding to Nvidia’s H100 chips in coaching LLMs however performs higher on the inference aspect — 1.4 instances higher than H100 when working with Meta’s Llama 2, a 70 billion parameter LLM.
AMD partnered with Microsoft to place MI300X in its Azure digital machines. Microsoft CTO Kevin Scott, a visitor throughout Su’s speech, additionally introduced the Azure ND MI300X digital machines — first revealed in November — are actually out there on preview. Meta additionally introduced it should deploy MI300 processors in its knowledge facilities.
Su stated AMD launched the MI300A APU for knowledge facilities, which she stated are anticipated to develop its complete addressable market to $45 billion. APUs usually mix CPUs and GPUs for quicker processing. AMD stated the MI300A presents higher-performance computing, quicker mannequin coaching, and a 30 instances vitality effectivity enchancment. In comparison with the H100, AMD stated it has 1.6 instances the reminiscence capability. It additionally options unified reminiscence, so there isn’t any want to maneuver knowledge from completely different gadgets anymore.
MI300A will energy the El Capitan supercomputer constructed by Hewlett Packard Enterprise on the Lawrence Livermore Nationwide Laboratory. El Capitan is taken into account some of the highly effective supercomputers and is anticipated to ship greater than two exaflops of efficiency.
The MI300A APU “is now in manufacturing and is being constructed into knowledge facilities.”
Pricing data was not instantly out there.
Su teased the MI300 chips throughout the Code Convention, saying AMD was excited concerning the alternative to faucet extra chip customers, not simply from cloud suppliers however from enterprises and startups.
AMD additionally introduced the most recent addition to its Ryzen processors, the Ryzen 8040, which might put extra native AI features into cellular gadgets. The corporate stated the 8040 collection presents 1.6 instances extra AI processing efficiency than earlier fashions and integrates neural processing items (NPUs).
The corporate stated Ryzen 8040 wouldn’t be restricted to AI processing, because it claimed video enhancing can be 65 % quicker and gaming can be 77 % quicker than with competing merchandise like Intel’s chips.
AMD expects producers like Acer, Asus, Dell, HP, Lenovo, and Razer to launch merchandise integrating Ryzen 8040 chips within the first quarter of 2024.
Su stated the subsequent era of its Strix Level NPUs can be launched in 2024.
AMD additionally introduced the Ryzen AI Software program Platform is now broadly out there, which can let builders constructing AI fashions on Ryzen-powered laptops offload fashions into the NPU so the CPU can cut back energy consumption. Customers will get help for basis fashions just like the speech recognition mannequin Whisper and LLMs like Llama 2.
To energy AI fashions — and reap the benefits of the present hype for the tech — firms like AMD, Nvidia, and Intel have been locked in what’s mainly an AI chip arms race. To this point, Nvidia captured the biggest market share with its extremely coveted H100 GPUs used to coach fashions like OpenAI’s GPT.