Skip to content

Commit 150e012

Browse files
committed
OSDOCS-14333: adds OTEL for AI model-server metrics MicroShift
1 parent 9ecb5b5 commit 150e012

4 files changed

+46
-2
lines changed

microshift_ai/microshift-rhoai.adoc

+2
Original file line numberDiff line numberDiff line change
@@ -64,6 +64,8 @@ include::modules/microshift-rhoai-servingruntimes-ex.adoc[leveloffset=+1]
6464
6565
include::modules/microshift-rhoai-inferenceservice-ex.adoc[leveloffset=+1]
6666

67+
include::modules/microshift-rhoai-inferenceservice-otel.adoc[leveloffset=+2]
68+
6769
include::modules/microshift-inferenceservice-more-options.adoc[leveloffset=+2]
6870

6971
include::modules/microshift-rhoai-model-serving-rt-verify.adoc[leveloffset=+1]

modules/microshift-inferenceservice-more-options.adoc

+2-2
Original file line numberDiff line numberDiff line change
@@ -2,9 +2,9 @@
22
//
33
// * microshift_ai/microshift-rhoai.adoc
44

5-
:_mod-docs-content-type: CONCEPT
5+
:_mod-docs-content-type: REFERENCE
66
[id="microshift-rhoai-inferenceservice-more-options_{context}"]
7-
= More InferenceService CRD options
7+
= More InferenceService CR options
88

99
The inference service YAML file can include many different options. For example, you can include a `resources` section that is passed first to the deployment and then to the pod, so that the model server gets access to your hardware through the device plugin.
1010

modules/microshift-rhoai-get-model-server-metrics.adoc

+5
Original file line numberDiff line numberDiff line change
@@ -8,6 +8,11 @@
88

99
After making a query, you can get the model server's metrics to identify bottlenecks, optimize resource allocation, and ensure efficient infrastructure utilization.
1010

11+
[NOTE]
12+
====
13+
You can alternatively configure Open Telemetry for {microshift-short} to get model-server metrics. See "Adding Open Telemetry to an InferenceService custom resource" for more information.
14+
====
15+
1116
.Prerequisites
1217

1318
* The {microshift-short} cluster is running.
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,37 @@
1+
// Module included in the following assemblies:
2+
//
3+
// * microshift_ai/microshift-rhoai.adoc
4+
5+
:_mod-docs-content-type: PROCEDURE
6+
[id="microshift-rhoai-inferenceservice-otel_{context}"]
7+
= Adding Open Telemetry to an InferenceService custom resource
8+
9+
You can get model metrics by using Open Telemetry if you installed the Open Telemetry RPM for {microshift-short}.
10+
11+
[NOTE]
12+
====
13+
You can alternatively get the Prometheus-format metrics of the model server by making a request on the `/metrics` endpoint. See "Getting the model-server metrics" for more information.
14+
====
15+
16+
.Prerequisites
17+
18+
* You configured the `ServingRuntimes` CR.
19+
* You have root user access to your machine.
20+
* The {oc-first} is installed.
21+
* You installed the `microshift-otel` RPM.
22+
* Your {microshift-short} Open Telemetry configuration includes the Prometheus Receiver. For more information, see link:https://docs.redhat.com/en/documentation/openshift_container_platform/4.18/html/red_hat_build_of_opentelemetry/configuring-the-collector#prometheus-receiver_otel-collector-receivers[Prometheus Receiver].
23+
24+
.Procedure
25+
26+
. Add the following Open Telemetry annotation to your `InferenceService` custom resource:
27+
+
28+
.Example `InferenceService` object with Open Telemetry
29+
[source,yaml]
30+
----
31+
apiVersion: serving.kserve.io/v1beta1
32+
kind: InferenceService
33+
metadata:
34+
name: ovms-resnet50
35+
annotations:
36+
prometheus.io/scrape: "true"
37+
----

0 commit comments

Comments
 (0)