This demo has been created for use with IBM Spyre Hardware on the Red Hat OpenShift platform.
The following automates the configuration of prerequisites and is specifically intended for
users with cluster-admin or an operations team.
User Model Serving listed below should be able to be run by a user with the self-provisioner role or access to create projects / namespaces.
Note
Verify your permissions if you experience errors.
- Red Hat OpenShift 4.11+
- Red Hat OpenShift AI 2.25+
- IBM AIU Spyre 1.0 (DD2)
- OpenShift Web Terminal (see below for installation instructions)
- OpenShift
cluster-adminpermissions
Install the OpenShift Web Terminal
Note
The following icon should appear in the top right of the OpenShift web console after you have installed the operator. Clicking this icon launches the web terminal.
NOTE: Reload the page in your browser if you do not see the icon after installing the operator.
If you do not wish to use the Quick Start, you may manually perform the demo installation steps using the following instructions.
Important
Run the following commands from the enhanced web terminal.
cluster-admin is required
# apply the enhanced (OpenShift) web terminal
oc apply -k https://github.com/redhat-na-ssa/demo-ibm-spyre/demo/web-terminal
# delete old web terminal
$(wtoctl | grep 'oc delete')Important
cluster-admin is required
These instructions will setup the operators
apply_firmly gitops/operatorsapply_firmly gitops/instance/model-servingThe following will setup model serving for a user with self-provisioner
# create a new project / namespace
oc new-project [username]
# create model inference
oc apply -k gitops/instance/model-serving/baseapply_firmly gitops/instance/rhoai-notebooksOnce the model is deployed, you can use a curl command to send a request to the model and perform inferencing. For example:
curl -k "http://granite-3-1-8b-instruct-predictor.demo-ibm-spyre.svc.cluster.local/v1/completions" \
-H "Content-Type: application/json" \
-d '{"model":"granite-3-1-8b-instruct","prompt":"Write a short poem.","temperature":0,"max_tokens":128}'oc delete -k gitops/instance/model-serving