|
| 1 | +// Module included in the following assemblies: |
| 2 | +// |
| 3 | +// * microshift_ai/microshift-rhoai.adoc |
| 4 | + |
| 5 | +:_mod-docs-content-type: CONCEPT |
| 6 | +[id="microshift-rhoai-install_{context}"] |
| 7 | += Installing the {rhoai-full} RPM |
| 8 | + |
| 9 | +To use AI models in {microshift-short} deployments, use this procedure to install the {rhoai-full} {{rhoai}} RPM a new {microshift-short} installation. You can install the RPM on an existing {microshift} instance as long as you restart the system. |
| 10 | + |
| 11 | +.Prerequisites |
| 12 | + |
| 13 | +* The system requirements for installing {microshift-short} have been met. |
| 14 | +* You have root user access to your machine. |
| 15 | +* You configured your LVM VG with the capacity needed for the PVs of your workload. |
| 16 | +//Do we need to config the CSI? |
| 17 | +* You added memory as required for your AI model. |
| 18 | +//I assume there is some kind of memory requirement for using an AI model? |
| 19 | +
|
| 20 | +
|
| 21 | +. Install the {microshift-short} AI-model-serving RPM package by running the following command: |
| 22 | ++ |
| 23 | +[source,terminal] |
| 24 | +---- |
| 25 | +$ sudo dnf install microshift-microshift-ai-model-serving |
| 26 | +---- |
| 27 | ++ |
| 28 | +[TIP] |
| 29 | +==== |
| 30 | +If you create your manifests and custom resources (CRs) while you are completing your installation of {microshift-short}, you can avoid restarting the service a second time to apply them. |
| 31 | +==== |
| 32 | + |
| 33 | +. As a root user, restart the {microshift-short} service by entering the following command: |
| 34 | ++ |
| 35 | +[source,terminal] |
| 36 | +---- |
| 37 | +$ sudo systemctl restart microshift |
| 38 | +---- |
| 39 | + |
| 40 | +//Q: Can we check the namespace to verify? |
| 41 | +//Kserve manifest and ServingRuntimes CRs are deployed in a `redhat-ods-applications` namespace. |
| 42 | + |
| 43 | +.Next steps |
| 44 | + |
| 45 | +. Configure your hardware, the operating system, and additional components to make use of their accelerators. |
| 46 | + |
| 47 | +. Create manifests that include |
| 48 | +* `ServingRuntime` in your namespace |
| 49 | +* `InferenceService` referencing `ServingRuntime` and a model |
| 50 | +* `Route` CR |
| 51 | + |
| 52 | +// Kserve creates Deployment and other resources. |
| 53 | + |
| 54 | +. Resources from previous step become ready and user can make HTTP/GRPC calls to the model server. |
0 commit comments