AMD ROCm 6.2

Introducing the Newest AMD ROCm 6.2 Release, Unleashing Next-Gen  AI & HPC Performance

AMD is adding vLLM support to AMD Instinct Accelerators to increase AI model performance and scalability

ROCm/vLLM offers FP8 GEMMs with bespoke decode paged attention for cutting-edge performance

New and existing AMD Instinct customers may simply integrate vLLM into their AI pipelines with ROCm 6.2

AI development is revolutionized by the Bits and bytes quantization library support via AMD ROCm

AI is optimized via LLM.Int8() quantization, enabling efficient LLM deployment on devices with lower memory

Bits and bytes democratizes AI development, gives cost savings, increases innovation potential, and makes advanced AI capabilities

Wide FP8 Support in ROCm can greatly enhance the AI model execution process, especially for inferencing

FP8’s lower precision calculations can reduce the latency associated with computations and data transfers

ROCm 6.2 supports FP8 in its frameworks, libraries, and more to improve performance and efficiency