The AMD Instinct MI300X is an HPC GPU from AMD designed for the AI market.
Part of the AMD Instinct family, the MI300X is a follow-on from the MI300A chip. Where the M300A combinedcombines three Zen4 CPU chiplets with multiple GPU chiplets, the MI300X replaces the CPUs with two additional CDNA 3 chiplets. This makes the MI300X design simpler than the MI300A with a 12 chiplets in total - 8total—8 GPU and 4 IO memory chiplets. The MI300X increases the transistor count from 146 billion transistors to 153 billion, and the shared DRAM memory is boosted from 128GB in the MI300A to 192GB. The memory bandwidth has increased from 800 gigabytes per second to 5.2 terabytes per second.
AMD first announced its MI300 series of chips in June 2022, and further details were released at CES in January 2023. The MI300X chip was introduced by CEO Lisa Su on June 13, 2023. AMD stated it will be sampling the MI300X GPU to customers in Q3 of 2023. As part of the presentation, Su stated the MI300X is powerful enough to run the Falcon-40B LLM (one of the most popular LLMs at the time with 40 billion parameters) entirely in memory, rather than moving data back and forth to external memory. Su said the MI300X could run models up to approximately 80 billion parameters in memory, helping to reduce:
The AMD Instinct MI300X is an HPC GPU from AMD designed for the AI market.
The AMD Instinct MI300X is a high-performance computing (HPC) graphics processing unit (GPU) from AMD designed for the AI market. The MI300X is a pure GPU, using CDNA 3 GPU tiles that are paired with 192GB of HBM3 memory. The chip is targeting large language model (LLM) developers that require significant memory capacity to run cutting-edge models.
Part of the AMD Instinct family, the MI300X is a follow-on from the MI300A chip. Where the M300A combined three Zen4 CPU chiplets with multiple GPU chiplets, the MI300X replaces the CPUs with two additional CDNA 3 chiplets. This makes the MI300X design simpler than the MI300A with a 12 chiplets in total - 8 GPU and 4 IO memory chiplets. The MI300X increases the transistor count from 146 billion transistors to 153 billion, and the shared DRAM memory is boosted from 128GB in the MI300A to 192GB. The memory bandwidth has increased from 800 gigabytes per second to 5.2 terabytes per second.
AMD first announced its MI300 series of chips in June 2022 and further details were released at CES in January 2023. The MI300X chip was introduced by CEO Lisa Su on June 13, 2023. AMD stated it will be sampling the MI300X GPU to customers in Q3 of 2023. As part of the presentation, Su stated the MI300X is powerful enough to run the Falcon-40B LLM (one of the most popular LLMs at the time with 40 billion parameters) entirely in memory, rather than moving data back and forth to external memory. Su said the MI300X could run models up to approximately 80 billion parameters in memory, helping to reduce:
the number of GPUs you need, significantly speeding up the performance, especially for inference, as well as reducing the total cost of ownership.
On June 13, 2023, AMD also announced the AMD Infinity Architecture Platform with an 8-way interlinked MI300X design for larger workloads.
June 13, 2023
2023
June 9, 2022
The announcement was part of an update on their CDNA server GPU architecture and related Instinct products.
The AMD Instinct MI300X is an HPC GPU from AMD designed for the AI market.