Skip to content

Commit e8d45f0

Browse files
committed
rename api and metrics
1 parent 6642dfa commit e8d45f0

File tree

6 files changed

+225
-230
lines changed

6 files changed

+225
-230
lines changed

pkg/ext-proc/metrics/metrics.go

+14-14
Original file line numberDiff line numberDiff line change
@@ -9,37 +9,37 @@ import (
99
)
1010

1111
const (
12-
LLMServiceModelComponent = "llmservice_model"
12+
InferenceModelComponent = "inference_model"
1313
)
1414

1515
var (
1616
requestCounter = compbasemetrics.NewCounterVec(
1717
&compbasemetrics.CounterOpts{
18-
Subsystem: LLMServiceModelComponent,
18+
Subsystem: InferenceModelComponent,
1919
Name: "request_total",
20-
Help: "Counter of LLM service requests broken out for each model and target model.",
20+
Help: "Counter of inference model requests broken out for each model and target model.",
2121
StabilityLevel: compbasemetrics.ALPHA,
2222
},
23-
[]string{"llmservice_name", "model_name", "target_model_name"},
23+
[]string{"model_name", "target_model_name"},
2424
)
2525

2626
requestLatencies = compbasemetrics.NewHistogramVec(
2727
&compbasemetrics.HistogramOpts{
28-
Subsystem: LLMServiceModelComponent,
28+
Subsystem: InferenceModelComponent,
2929
Name: "request_duration_seconds",
30-
Help: "LLM service response latency distribution in seconds for each model and target model.",
30+
Help: "Inference model response latency distribution in seconds for each model and target model.",
3131
Buckets: []float64{0.005, 0.025, 0.05, 0.1, 0.2, 0.4, 0.6, 0.8, 1.0, 1.25, 1.5, 2, 3,
3232
4, 5, 6, 8, 10, 15, 20, 30, 45, 60, 120, 180, 240, 300, 360, 480, 600, 900, 1200, 1800, 2700, 3600},
3333
StabilityLevel: compbasemetrics.ALPHA,
3434
},
35-
[]string{"llmservice_name", "model_name", "target_model_name"},
35+
[]string{"model_name", "target_model_name"},
3636
)
3737

3838
requestSizes = compbasemetrics.NewHistogramVec(
3939
&compbasemetrics.HistogramOpts{
40-
Subsystem: LLMServiceModelComponent,
40+
Subsystem: InferenceModelComponent,
4141
Name: "request_sizes",
42-
Help: "LLM service requests size distribution in bytes for each model and target model.",
42+
Help: "Inference model requests size distribution in bytes for each model and target model.",
4343
// Use buckets ranging from 1000 bytes (1KB) to 10^9 bytes (1GB).
4444
Buckets: []float64{
4545
64, 128, 256, 512, 1024, 2048, 4096, 8192, 16384, 32768, 65536, // More fine-grained up to 64KB
@@ -48,7 +48,7 @@ var (
4848
},
4949
StabilityLevel: compbasemetrics.ALPHA,
5050
},
51-
[]string{"llmservice_name", "model_name", "target_model_name"},
51+
[]string{"model_name", "target_model_name"},
5252
)
5353
)
5454

@@ -64,9 +64,9 @@ func Register() {
6464
}
6565

6666
// MonitorRequest handles monitoring requests.
67-
func MonitorRequest(llmserviceName, modelName, targetModelName string, reqSize int, elapsed time.Duration) {
67+
func MonitorRequest(modelName, targetModelName string, reqSize int, elapsed time.Duration) {
6868
elapsedSeconds := elapsed.Seconds()
69-
requestCounter.WithLabelValues(llmserviceName, modelName, targetModelName).Inc()
70-
requestLatencies.WithLabelValues(llmserviceName, modelName, targetModelName).Observe(elapsedSeconds)
71-
requestSizes.WithLabelValues(llmserviceName, modelName, targetModelName).Observe(float64(reqSize))
69+
requestCounter.WithLabelValues(modelName, targetModelName).Inc()
70+
requestLatencies.WithLabelValues(modelName, targetModelName).Observe(elapsedSeconds)
71+
requestSizes.WithLabelValues(modelName, targetModelName).Observe(float64(reqSize))
7272
}

pkg/ext-proc/metrics/metrics_test.go

+4-9
Original file line numberDiff line numberDiff line change
@@ -9,13 +9,12 @@ import (
99
"k8s.io/component-base/metrics/testutil"
1010
)
1111

12-
const RequestTotalMetric = LLMServiceModelComponent + "_request_total"
13-
const RequestLatenciesMetric = LLMServiceModelComponent + "_request_duration_seconds"
14-
const RequestSizesMetric = LLMServiceModelComponent + "_request_sizes"
12+
const RequestTotalMetric = InferenceModelComponent + "_request_total"
13+
const RequestLatenciesMetric = InferenceModelComponent + "_request_duration_seconds"
14+
const RequestSizesMetric = InferenceModelComponent + "_request_sizes"
1515

1616
func TestMonitorRequest(t *testing.T) {
1717
type requests struct {
18-
llmserviceName string
1918
modelName string
2019
targetModelName string
2120
reqSize int
@@ -28,28 +27,24 @@ func TestMonitorRequest(t *testing.T) {
2827
name: "multiple requests",
2928
reqs: []requests{
3029
{
31-
llmserviceName: "s10",
3230
modelName: "m10",
3331
targetModelName: "t10",
3432
reqSize: 1200,
3533
elapsed: time.Millisecond * 10,
3634
},
3735
{
38-
llmserviceName: "s10",
3936
modelName: "m10",
4037
targetModelName: "t10",
4138
reqSize: 500,
4239
elapsed: time.Millisecond * 1600,
4340
},
4441
{
45-
llmserviceName: "s10",
4642
modelName: "m10",
4743
targetModelName: "t11",
4844
reqSize: 2480,
4945
elapsed: time.Millisecond * 60,
5046
},
5147
{
52-
llmserviceName: "s20",
5348
modelName: "m20",
5449
targetModelName: "t20",
5550
reqSize: 80,
@@ -61,7 +56,7 @@ func TestMonitorRequest(t *testing.T) {
6156
for _, scenario := range scenarios {
6257
t.Run(scenario.name, func(t *testing.T) {
6358
for _, req := range scenario.reqs {
64-
MonitorRequest(req.llmserviceName, req.modelName, req.targetModelName, req.reqSize, req.elapsed)
59+
MonitorRequest(req.modelName, req.targetModelName, req.reqSize, req.elapsed)
6560
}
6661
wantRequestTotal, err := os.Open("testdata/request_total_metric")
6762
defer func() {

pkg/ext-proc/metrics/testdata/request_duration_seconds_metic

Whitespace-only changes.

0 commit comments

Comments
 (0)