AMDInstinct

AMD Instinct Accelerators

AMD Instinct delivers industry-leading memory capacity for enterprise AI. MI300X offers 192GB HBM3 on CDNA 3, MI325X extends to 256GB HBM3E with optimized performance, and MI355X brings CDNA 4 architecture with 288GB HBM3E and 8TB/s bandwidth. The competitive alternative for large-scale AI training and inference.

MI300X
CDNA 3
192GB
MI325X
CDNA 3
256GB
CDNA 4
MI355X
Next Gen
288GB
192-288GB HBM3/HBM3E Memory
8TB/s Max Bandwidth
20 PFLOPS FP4 (MI355X)
ROCm Software Stack

Why AMD Instinct

Competitive performance with industry-leading memory capacity

288GB
Memory Leadership
Largest HBM3E capacity available
8TB/s
Memory Bandwidth
MI355X CDNA 4 architecture
Open
ROCm Ecosystem
PyTorch, TensorFlow native support
Value
Competitive Pricing
More memory per dollar

AMD Instinct Lineup

Choose the right Instinct accelerator for your workload

CDNA 3 MI300X
Memory 192GB HBM3
Bandwidth 5.3TB/s
FP32 163.4 TFLOPS
TDP 750W
Enterprise Price $10,000 - $15,000
Available now · Proven performance
CDNA 3 MI325X
Memory 256GB HBM3E
Bandwidth 6TB/s
FP16 1307 TFLOPS
TDP ~800W
Enterprise Price $15,000 - $20,000
Available now · More memory + bandwidth

Technical Specifications

Complete AMD Instinct technical details

Specification MI300X MI325X MI355X
Architecture
GPU Architecture CDNA 3 CDNA 3 CDNA 4
Process Node TSMC 5nm / 6nm TSMC 3nm
Compute Units 304 CUs 256 CUs
Stream Processors 19,456 16,384
Memory
Memory Capacity 192GB HBM3 256GB HBM3E 288GB HBM3E
Memory Bandwidth 5.3TB/s 6TB/s 8TB/s
Memory Bus Width 8192-bit
L2 Cache 256MB TBD
ECC Memory Yes
Performance
FP64 81.7 TFLOPS 81.7 TFLOPS 78.6 TFLOPS
FP32 163.4 TFLOPS 163.4 TFLOPS 157.2 TFLOPS
FP16 1307 TFLOPS 1307 TFLOPS 5 PFLOPS
FP8 2614 TFLOPS 2614 TFLOPS 10.1 PFLOPS
FP6/FP4 N/A N/A 20.1 PFLOPS
Connectivity
Infinity Fabric 7 links @ 64GB/s (448GB/s) Enhanced IF
PCIe PCIe 5.0 x16
Form Factor OAM
Power & Thermal
TDP 750W ~800W 1400W
Cooling Liquid Required
Software
Software Stack ROCm 6.0+
Framework Support PyTorch, TensorFlow
Virtualization SR-IOV

Ideal Use Cases

AMD Instinct excels at memory-intensive AI workloads

Large Language Models

Industry-leading memory capacity (up to 288GB) enables running larger LLMs without model parallelism overhead.

AI Training

Massive memory bandwidth (up to 8TB/s) accelerates training of foundation models without memory bottlenecks.

High-Throughput Inference

MI355X's FP4 support delivers 20+ PFLOPS for efficient large-scale inference deployments.

Scientific Computing

Strong FP64 performance (~80 TFLOPS) for HPC workloads including climate simulation and physics.

Memory-Bound Workloads

8TB/s bandwidth eliminates memory bottlenecks for data-intensive analytics and graph processing.

Open Source AI

ROCm provides open-source software stack with native PyTorch and TensorFlow support for flexibility.

Architecture Comparison

How AMD Instinct compares to NVIDIA data center GPUs

MI355X vs B200

  • Memory: 288GB vs 192GB
  • Bandwidth: 8TB/s vs 8TB/s
  • FP4: 20.1 PFLOPS vs 20 PFLOPS
  • TDP: 1400W vs 1000W
  • Ecosystem: ROCm vs CUDA
  • Price: Often lower
Verdict: Competitive performance; choose based on ecosystem preference

MI300X vs H100

  • Memory: 192GB vs 80GB (+140%)
  • Bandwidth: 5.3TB/s vs 3.35TB/s (+58%)
  • FP32: 163.4 vs 67 TFLOPS (+144%)
  • TDP: 750W vs 700W
  • Ecosystem: ROCm vs CUDA
  • Price: $10-15K vs $35-40K
Verdict: MI300X offers more memory/value; H100 has mature CUDA ecosystem

Frequently Asked Questions

What is the difference between MI300X, MI325X, and MI355X?

MI300X (CDNA 3): 192GB HBM3, 5.3TB/s bandwidth, 750W - the proven baseline. MI325X (CDNA 3): 256GB HBM3E, 6TB/s bandwidth with tuned performance. MI355X (CDNA 4): 288GB HBM3E, 8TB/s bandwidth, 1400W with new FP6/FP4 support delivering 20+ PFLOPS for AI inference.

How does AMD Instinct compare to NVIDIA?

AMD Instinct offers competitive or superior memory capacity (192-288GB) and bandwidth at often lower price points. NVIDIA has the more mature CUDA ecosystem while AMD uses ROCm with native PyTorch/TensorFlow support. For memory-intensive workloads, AMD often provides better value.

What software stack does AMD Instinct use?

AMD Instinct uses ROCm (Radeon Open Compute) software stack version 6.0+ with native support for PyTorch and TensorFlow. ROCm provides a CUDA-compatible programming model through HIP, though some CUDA code may require modifications for optimal performance.

When is MI355X available?

MI355X reached General Availability (GA) in October 2025. It's available through major OEM partners including Dell, HPE, Supermicro, Lenovo, and cloud providers like Oracle Cloud Infrastructure. Contact SLYD for current availability.

What cooling does AMD Instinct require?

All AMD Instinct accelerators require liquid cooling infrastructure. MI300X has 750W TDP, MI325X has 1000W TDP, and MI355X has 1400W peak board power. The OAM form factor is designed for high-performance server platforms with integrated liquid cooling.

Should I choose AMD Instinct or NVIDIA?

Choose AMD Instinct if you need maximum memory capacity, value competitive pricing, or prefer open-source software (ROCm). Choose NVIDIA if you need maximum CUDA ecosystem compatibility, specific NVIDIA features, or proven enterprise support. Both deliver excellent performance for AI workloads.

Deploy AMD Instinct

Get enterprise AMD Instinct systems through our OEM partnerships with expert configuration and deployment support.

Available through:
Dell HPE Supermicro Lenovo Gigabyte
Reconnecting to the server...

Please wait while we restore your connection

An unhandled error has occurred. Reload 🗙