|
| 1 | +# HELP inference_model_input_tokens [ALPHA] Inference model input token count distribution for requests in each model. |
| 2 | +# TYPE inference_model_input_tokens histogram |
| 3 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t10",le="1"} 0 |
| 4 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t10",le="8"} 0 |
| 5 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t10",le="16"} 1 |
| 6 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t10",le="32"} 2 |
| 7 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t10",le="64"} 2 |
| 8 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t10",le="128"} 2 |
| 9 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t10",le="256"} 2 |
| 10 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t10",le="512"} 2 |
| 11 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t10",le="1024"} 2 |
| 12 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t10",le="2048"} 2 |
| 13 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t10",le="4096"} 2 |
| 14 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t10",le="8192"} 2 |
| 15 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t10",le="16384"} 2 |
| 16 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t10",le="32778"} 2 |
| 17 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t10",le="65536"} 2 |
| 18 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t10",le="131072"} 2 |
| 19 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t10",le="262144"} 2 |
| 20 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t10",le="524288"} 2 |
| 21 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t10",le="1.048576e+06"} 2 |
| 22 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t10",le="+Inf"} 2 |
| 23 | +inference_model_input_tokens_sum{model_name="m10",target_model_name="t10"} 30 |
| 24 | +inference_model_input_tokens_count{model_name="m10",target_model_name="t10"} 2 |
| 25 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t11",le="1"} 0 |
| 26 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t11",le="8"} 0 |
| 27 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t11",le="16"} 0 |
| 28 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t11",le="32"} 1 |
| 29 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t11",le="64"} 1 |
| 30 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t11",le="128"} 1 |
| 31 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t11",le="256"} 1 |
| 32 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t11",le="512"} 1 |
| 33 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t11",le="1024"} 1 |
| 34 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t11",le="2048"} 1 |
| 35 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t11",le="4096"} 1 |
| 36 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t11",le="8192"} 1 |
| 37 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t11",le="16384"} 1 |
| 38 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t11",le="32778"} 1 |
| 39 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t11",le="65536"} 1 |
| 40 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t11",le="131072"} 1 |
| 41 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t11",le="262144"} 1 |
| 42 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t11",le="524288"} 1 |
| 43 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t11",le="1.048576e+06"} 1 |
| 44 | +inference_model_input_tokens_bucket{model_name="m10",target_model_name="t11",le="+Inf"} 1 |
| 45 | +inference_model_input_tokens_sum{model_name="m10",target_model_name="t11"} 30 |
| 46 | +inference_model_input_tokens_count{model_name="m10",target_model_name="t11"} 1 |
| 47 | +inference_model_input_tokens_bucket{model_name="m20",target_model_name="t20",le="1"} 0 |
| 48 | +inference_model_input_tokens_bucket{model_name="m20",target_model_name="t20",le="8"} 0 |
| 49 | +inference_model_input_tokens_bucket{model_name="m20",target_model_name="t20",le="16"} 0 |
| 50 | +inference_model_input_tokens_bucket{model_name="m20",target_model_name="t20",le="32"} 0 |
| 51 | +inference_model_input_tokens_bucket{model_name="m20",target_model_name="t20",le="64"} 1 |
| 52 | +inference_model_input_tokens_bucket{model_name="m20",target_model_name="t20",le="128"} 1 |
| 53 | +inference_model_input_tokens_bucket{model_name="m20",target_model_name="t20",le="256"} 1 |
| 54 | +inference_model_input_tokens_bucket{model_name="m20",target_model_name="t20",le="512"} 1 |
| 55 | +inference_model_input_tokens_bucket{model_name="m20",target_model_name="t20",le="1024"} 1 |
| 56 | +inference_model_input_tokens_bucket{model_name="m20",target_model_name="t20",le="2048"} 1 |
| 57 | +inference_model_input_tokens_bucket{model_name="m20",target_model_name="t20",le="4096"} 1 |
| 58 | +inference_model_input_tokens_bucket{model_name="m20",target_model_name="t20",le="8192"} 1 |
| 59 | +inference_model_input_tokens_bucket{model_name="m20",target_model_name="t20",le="16384"} 1 |
| 60 | +inference_model_input_tokens_bucket{model_name="m20",target_model_name="t20",le="32778"} 1 |
| 61 | +inference_model_input_tokens_bucket{model_name="m20",target_model_name="t20",le="65536"} 1 |
| 62 | +inference_model_input_tokens_bucket{model_name="m20",target_model_name="t20",le="131072"} 1 |
| 63 | +inference_model_input_tokens_bucket{model_name="m20",target_model_name="t20",le="262144"} 1 |
| 64 | +inference_model_input_tokens_bucket{model_name="m20",target_model_name="t20",le="524288"} 1 |
| 65 | +inference_model_input_tokens_bucket{model_name="m20",target_model_name="t20",le="1.048576e+06"} 1 |
| 66 | +inference_model_input_tokens_bucket{model_name="m20",target_model_name="t20",le="+Inf"} 1 |
| 67 | +inference_model_input_tokens_sum{model_name="m20",target_model_name="t20"} 40 |
| 68 | +inference_model_input_tokens_count{model_name="m20",target_model_name="t20"} 1 |
0 commit comments