logo80lv
Articlesclick_arrow
Research
Talentsclick_arrow
Events
Workshops
Aboutclick_arrow
profile_loginLogIn

AMD Announced MI300X Accelerator for Generative AI

It is based on the next-gen AMD CDNA 3 accelerator architecture and supports up to 192 GB of HBM3 memory.

NVIDIA has been expanding its AI horizons for some time now, so it's time for AMD to offer something, too. On Tuesday, the company announced an AMD Instinct MI300X accelerator, "the world’s most advanced accelerator for generative AI."

It is based on the next-gen AMD CDNA 3 accelerator architecture and supports up to 192 GB of HBM3 memory to provide the power needed for large language model training and inference for generative AI workloads. AMD also introduced the AMD Instinct Platform, bringing together eight MI300X accelerators "for the ultimate solution for AI inference and training."

"AI is the defining technology shaping the next generation of computing and the largest strategic growth opportunity for AMD," said AMD Chair and CEO Dr. Lisa Su. "We are laser focused on accelerating the deployment of AMD AI platforms at scale in the data center, led by the launch of our Instinct MI300 accelerators planned for later this year and the growing ecosystem of enterprise-ready AI software optimized for our hardware."

The MI300X will be available to key customers starting in Q3. Moreover, AMD's Instinct MI300A, an APU accelerator for HPC and AI workloads, is now sampling to customers.

In addition, AMD showcased the ROCm software ecosystem for data center accelerators. It will get “day zero” support for PyTorch 2.0 with ROCm release 5.4.2 on all AMD Instinct accelerators. Hugging Face will optimize thousands of its models on AMD platforms, "from AMD Instinct accelerators to AMD Ryzen and AMD EPYC processors, AMD Radeon GPUs and Versal and Alveo adaptive processors."

AMD also released a series of updates to its 4th Gen EPYC family:

  • Advancing the World’s Best Data Center CPU. AMD highlighted how the 4th Gen AMD EPYC processor continues to drive leadership performance and energy efficiency. AMD was joined by AWS to highlight a preview of the next generation Amazon Elastic Compute Cloud (Amazon EC2) M7a instances, powered by 4th Gen AMD EPYC processors (“Genoa”). Outside of the event, Oracle announced plans to make available new Oracle Computing Infrastructure (OCI) E5 instances with 4th Gen AMD EPYC processors.
  • No Compromise Cloud Native Computing. AMD introduced the 4th Gen AMD EPYC 97X4 processors, formerly codenamed “Bergamo.” With 128 “Zen 4c” cores per socket, these processors provide the greatest vCPU density1 and industry leading2 performance for applications that run in the cloud, and leadership energy efficiency. AMD was joined by Meta who discussed how these processors are well suited for their mainstay applications such as Instagram, WhatsApp and more; how Meta is seeing impressive performance gains with 4th Gen AMD EPYC 97x4 processors compared to 3rd Gen AMD EPYC across various workloads, while offering substantial TCO improvements as well, and how AMD and Meta optimized the EPYC CPUs for Meta’s power-efficiency and compute-density requirements.
  • Enabling Better Products With Technical Computing. AMD introduced the 4th Gen AMD EPYC processors with AMD 3D V-Cache™ technology, the world’s highest performance x86 server CPU for technical computing3. Microsoft announced the general availability of Azure HBv4 and HX instances, powered by 4th Gen AMD EPYC processors with AMD 3D V-Cache technology.

Learn more here and don't forget to join our 80 Level Talent platform and our Telegram channel, follow us on Instagram and Twitter, where we share breakdowns, the latest news, awesome artworks, and more.

Join discussion

Comments 0

    You might also like

    We need your consent

    We use cookies on this website to make your browsing experience better. By using the site you agree to our use of cookies.Learn more

    ×