AMD GPU for HPC and AI¶
Timeline: 6–12 months.
Prerequisites: Phase 1 §4 (C++ and Parallel Computing — CUDA/OpenCL), Phase 4 Track C (ML compiler fundamentals), Nvidia GPU sub-track recommended for context.
Why AMD GPU¶
AMD's Instinct GPUs (MI300X, MI300A, MI350) are the primary alternative to Nvidia for AI training and inference at scale. Understanding both ecosystems makes you more versatile and valuable — especially as cloud providers (Azure, Oracle, Meta) deploy AMD hardware alongside Nvidia.
1. AMD GPU Architecture¶
- CDNA vs RDNA:
- CDNA (Compute DNA): data-center optimized — MI300X, MI250X. Matrix cores, large HBM, high bandwidth.
- RDNA (Radeon DNA): consumer/gaming GPUs. Relevant for understanding the architecture family.
- MI300X architecture:
- Chiplet design: 8 XCDs (Accelerator Complex Dies) + 4 IODs.
- 192 GB HBM3 with 5.3 TB/s bandwidth.
- Matrix cores: FP16, BF16, FP8, INT8.
- Infinity Fabric for inter-chiplet and inter-GPU communication.
- Compute units:
- Wavefront (64 threads) vs CUDA warp (32 threads).
- SIMD units, LDS (Local Data Share) vs CUDA shared memory.
- Occupancy model differences from Nvidia.
2. ROCm Software Stack¶
- ROCm (Radeon Open Compute):
- Open-source GPU compute platform. ROCm 6+ for MI300X.
- Kernel driver (
amdgpu), runtime (hip-runtime), compiler (amd-clang).
- HIP (Heterogeneous-computing Interface for Portability):
- CUDA-like API for AMD GPUs.
hipMalloc,hipMemcpy,hipLaunchKernelGGL. - HIPIFY: Tool to convert CUDA source to HIP (
hipify-perl,hipify-clang). - Write-once kernels that target both AMD and Nvidia GPUs.
- CUDA-like API for AMD GPUs.
- Libraries:
- rocBLAS (GEMM), MIOpen (cuDNN equivalent), rocFFT, rocSPARSE.
- RCCL (ROCm Communication Collectives Library) — AMD's NCCL equivalent for multi-GPU.
- Composable Kernel (CK) — AMD's equivalent to CUTLASS for custom GEMM/attention kernels.
- Profiling:
- rocProfiler / rocTracer — kernel-level profiling.
- Omniperf — roofline analysis, occupancy, memory throughput (like Nsight Compute).
- Omnitrace — timeline profiling (like Nsight Systems).
3. Porting CUDA to AMD¶
- HIPIFY workflow:
- Automated conversion:
hipify-clangfor source-to-source translation. - Manual work: CUDA-specific intrinsics, inline PTX, warp-level primitives.
- Automated conversion:
- Key differences:
- Warp size 32 (Nvidia) vs wavefront size 64 (AMD) — affects reduction, ballot, shuffle ops.
- Shared memory bank conflicts: 32 banks (Nvidia) vs 64 banks (AMD).
- Memory coalescing rules differ slightly.
- No direct equivalent to Nvidia Tensor Cores — use Matrix Cores via
rocWMMAor CK.
- Framework support:
- PyTorch: native ROCm support (
torch.cudaworks on AMD via HIP). - TensorFlow, JAX: ROCm backends available.
- TVM, ONNX Runtime: ROCm execution providers.
- vLLM, TensorRT-LLM alternatives for AMD: vLLM + ROCm, AMD Inference Server.
- PyTorch: native ROCm support (
4. Multi-GPU and Cluster Operations¶
- Infinity Fabric:
- Intra-node GPU-to-GPU interconnect. Bandwidth and topology compared to NVLink.
- MI300X: 896 GB/s aggregate fabric bandwidth per GPU.
- RCCL:
- All-reduce, all-gather, reduce-scatter on AMD GPUs.
- Tuning for MI300X topology and Infinity Fabric.
- Multi-node:
- InfiniBand and RoCE (RDMA over Converged Ethernet) support.
- GPUDirect RDMA equivalent on AMD.
- Slurm and Kubernetes with ROCm container support.
5. Kernel Development on AMD¶
- HIP kernel writing:
- Thread hierarchy: grid → block → thread (same as CUDA).
- Shared memory (
__shared__), synchronization (__syncthreads()). - Wavefront-aware programming:
__ballot,__shfl, warp intrinsics via wavefront equivalents.
- Composable Kernel (CK):
- Templated library for high-performance GEMM, attention, and custom ops.
- Tile-based programming model (similar concept to CuTe/CUTLASS).
- Writing custom kernels with CK building blocks.
- Triton on AMD:
- Triton supports AMD GPUs via ROCm backend.
- Same Python kernel code, different backend code generation.
- Performance tuning differences (tile sizes, occupancy targets).
Resources¶
Projects¶
- HIPIFY a CUDA kernel — Take your CUDA matmul or vector-add kernel from Phase 1 §4. Convert to HIP using
hipify-clang. Run on AMD GPU (or ROCm Docker). Compare output and performance. - Profile with Omniperf — Profile a PyTorch model on ROCm. Generate a roofline plot. Identify compute-bound vs memory-bound layers.
- RCCL benchmark — Run RCCL all-reduce across multiple AMD GPUs. Compare bandwidth and latency with NCCL on equivalent Nvidia hardware.
- Triton on AMD — Write a Triton fused kernel (e.g., layer norm + residual). Run on both Nvidia (CUDA) and AMD (ROCm). Compare generated code and performance.
- CK custom GEMM — Use Composable Kernel to implement a custom GEMM with epilogue fusion (bias + activation). Benchmark against rocBLAS.