IBM and AMD have announced a collaboration to deploy AMD Instinct MI300X accelerators as a service on IBM Cloud. This offering, which is expected to be available in the first half of 2025, will enhance performance and power efficiency for Generative AI models such as and high-performance computing (HPC) applications for enterprise clients.
To help enterprise clients across industries, including those that are heavily regulated, IBM and AMD intend to leverage IBM Cloud’s security and compliance capabilities below:
- Support for Large Model Inferencing: Equipped with 192GB of high-bandwidth memory (HBM3), AMD Instinct MI300X accelerators offer support for the largest model inferencing and fine tuning. The large memory capacity can also help customers run larger models with fewer GPUs, potentially lowering costs for inferencing.
- Enhanced Performance and Security: Offering AMD Instinct MI300X accelerators as a service on IBM Cloud Virtual Servers for VPC, as well as through container support with IBM Cloud Kubernetes Service and Red Hat OpenShift on IBM Cloud, can help optimize performance for enterprises running AI applications.
Read our newsroom article to learn more around these new AI accelerators.