As most already know, Red Hat OpenShift AI provides much of the underlying capabilities found in IBM's watsonx.ai. If your needs are best met by OpenShift AI and you want to quickly create a Red Hat OpenShift AI deployment in your IBM Cloud account, this blog is for you.
IBM Cloud has the concept of Deployable Architectures which describes them exactly, an architecture you can deploy. In our case we want a ROKS cluster that has the OpenShift AI stack of operators installed and ready for use. We can use a deployable architecture to quickly produce this environment for us. The good news is that you can create a Private Catalog where you can install a Deployable Architecture. We will leave that exercise for another blog post, but we can go over how to execute our Deployable Architecture catalog entry once you get it into your catalog.
First, you can find this Deployable Architecture here. You will notice it is mostly standard Terraform with a few extra files. You are free to execute that Terraform in your own environment, but let's look at how you would execute this Deployable Architecture.
In your catalog, we first have to find our Deployable Architecture tile. You can narrow them down by a filter:
You should now be able to find the tile we are looking for:
Click on it. You will see an overview page. At the bottom is a picture of what we will be creating with this Deployable Architecture. We will get a new VPC with a new subnet and attached public gateway. In that VPC we will create a ROKS cluster with GPU based worker nodes and finally the OpenShift AI operator and its dependencies will be installed. When complete you will have a cluster ready for your OpenShift AI exploration.