Skip to content

Commit 020c547

Browse files
authored
Merge pull request #92556 from openshift-cherrypick-robot/cherry-pick-92392-to-enterprise-4.19
[enterprise-4.19] OSDOCS-14344: adds OTEL for AI model-server metrics MicroShift
2 parents 40b306b + 5177143 commit 020c547

4 files changed

+48
-2
lines changed

microshift_ai/microshift-rhoai.adoc

+2
Original file line numberDiff line numberDiff line change
@@ -64,6 +64,8 @@ include::modules/microshift-rhoai-servingruntimes-ex.adoc[leveloffset=+1]
6464
6565
include::modules/microshift-rhoai-inferenceservice-ex.adoc[leveloffset=+1]
6666

67+
include::modules/microshift-rhoai-export-metrics-otel.adoc[leveloffset=+2]
68+
6769
include::modules/microshift-inferenceservice-more-options.adoc[leveloffset=+2]
6870

6971
include::modules/microshift-rhoai-model-serving-rt-verify.adoc[leveloffset=+1]

modules/microshift-inferenceservice-more-options.adoc

+2-2
Original file line numberDiff line numberDiff line change
@@ -2,9 +2,9 @@
22
//
33
// * microshift_ai/microshift-rhoai.adoc
44

5-
:_mod-docs-content-type: CONCEPT
5+
:_mod-docs-content-type: REFERENCE
66
[id="microshift-rhoai-inferenceservice-more-options_{context}"]
7-
= More InferenceService CRD options
7+
= More InferenceService CR options
88

99
The inference service YAML file can include many different options. For example, you can include a `resources` section that is passed first to the deployment and then to the pod, so that the model server gets access to your hardware through the device plugin.
1010

Original file line numberDiff line numberDiff line change
@@ -0,0 +1,39 @@
1+
// Module included in the following assemblies:
2+
//
3+
// * microshift_ai/microshift-rhoai.adoc
4+
5+
:_mod-docs-content-type: PROCEDURE
6+
[id="microshift-rhoai-export-metrics-otel_{context}"]
7+
= Exporting model-server metrics by using Open Telemetry
8+
9+
You can export model-server metrics by using Open Telemetry if you installed the `microshift-observability` RPM for {microshift-short}.
10+
11+
[NOTE]
12+
====
13+
You can alternatively get the Prometheus-format metrics of the model server by making a request on the `/metrics` endpoint. See "Getting the model-server metrics" for more information.
14+
====
15+
16+
.Prerequisites
17+
18+
* You configured the `ServingRuntimes` CR.
19+
* You have root user access to your machine.
20+
* The {oc-first} is installed.
21+
* You installed the `microshift-observability` RPM.
22+
* Your {microshift-short} Open Telemetry configuration includes the Prometheus Receiver. For more information, see link:https://docs.redhat.com/en/documentation/openshift_container_platform/4.18/html/red_hat_build_of_opentelemetry/configuring-the-collector#prometheus-receiver_otel-collector-receivers[Prometheus Receiver].
23+
24+
.Procedure
25+
26+
* Add the following Open Telemetry annotation to your `InferenceService` custom resource:
27+
+
28+
.Example `InferenceService` object with Open Telemetry
29+
[source,yaml]
30+
----
31+
apiVersion: serving.kserve.io/v1beta1
32+
kind: InferenceService
33+
metadata:
34+
name: ovms-resnet50
35+
#...
36+
annotations:
37+
prometheus.io/scrape: "true"
38+
#...
39+
----

modules/microshift-rhoai-get-model-server-metrics.adoc

+5
Original file line numberDiff line numberDiff line change
@@ -8,6 +8,11 @@
88

99
After making a query, you can get the model server's metrics to identify bottlenecks, optimize resource allocation, and ensure efficient infrastructure utilization.
1010

11+
[NOTE]
12+
====
13+
You can alternatively configure Open Telemetry for {microshift-short} to get model-server metrics. See "Adding Open Telemetry to an InferenceService custom resource" for more information.
14+
====
15+
1116
.Prerequisites
1217

1318
* The {microshift-short} cluster is running.

0 commit comments

Comments
 (0)