Skip to content

Commit e045668

Browse files
committed
Update test
1 parent 94426cc commit e045668

File tree

1 file changed

+33
-12
lines changed

1 file changed

+33
-12
lines changed

ci/L0_backend_vllm/metrics_test/vllm_metrics_test.py

Lines changed: 33 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -119,39 +119,60 @@ def test_vllm_metrics(self):
119119
model_name=self.vllm_model_name,
120120
)
121121
metrics_dict = self.parse_vllm_metrics()
122+
total_prompts = len(self.prompts)
122123

123124
# vllm:prompt_tokens_total
124125
self.assertEqual(metrics_dict["vllm:prompt_tokens_total"], 18)
125126
# vllm:generation_tokens_total
126127
self.assertEqual(metrics_dict["vllm:generation_tokens_total"], 48)
127128
# vllm:time_to_first_token_seconds
128-
self.assertEqual(metrics_dict["vllm:time_to_first_token_seconds_count"], 3)
129+
self.assertEqual(
130+
metrics_dict["vllm:time_to_first_token_seconds_count"], total_prompts
131+
)
129132
self.assertGreater(metrics_dict["vllm:time_to_first_token_seconds_sum"], 0)
130-
self.assertEqual(metrics_dict["vllm:time_to_first_token_seconds_bucket"], 3)
133+
self.assertEqual(
134+
metrics_dict["vllm:time_to_first_token_seconds_bucket"], total_prompts
135+
)
131136
# vllm:time_per_output_token_seconds
132137
self.assertEqual(metrics_dict["vllm:time_per_output_token_seconds_count"], 45)
133138
self.assertGreater(metrics_dict["vllm:time_per_output_token_seconds_sum"], 0)
134139
self.assertEqual(metrics_dict["vllm:time_per_output_token_seconds_bucket"], 45)
135140
# vllm:e2e_request_latency_seconds
136-
self.assertEqual(metrics_dict["vllm:e2e_request_latency_seconds_count"], 3)
141+
self.assertEqual(
142+
metrics_dict["vllm:e2e_request_latency_seconds_count"], total_prompts
143+
)
137144
self.assertGreater(metrics_dict["vllm:e2e_request_latency_seconds_sum"], 0)
138-
self.assertEqual(metrics_dict["vllm:e2e_request_latency_seconds_bucket"], 3)
145+
self.assertEqual(
146+
metrics_dict["vllm:e2e_request_latency_seconds_bucket"], total_prompts
147+
)
139148
# vllm:request_prompt_tokens
140-
self.assertEqual(metrics_dict["vllm:request_prompt_tokens_count"], 3)
149+
self.assertEqual(
150+
metrics_dict["vllm:request_prompt_tokens_count"], total_prompts
151+
)
141152
self.assertEqual(metrics_dict["vllm:request_prompt_tokens_sum"], 18)
142-
self.assertEqual(metrics_dict["vllm:request_prompt_tokens_bucket"], 3)
153+
self.assertEqual(
154+
metrics_dict["vllm:request_prompt_tokens_bucket"], total_prompts
155+
)
143156
# vllm:request_generation_tokens
144-
self.assertEqual(metrics_dict["vllm:request_generation_tokens_count"], 3)
157+
self.assertEqual(
158+
metrics_dict["vllm:request_generation_tokens_count"], total_prompts
159+
)
145160
self.assertEqual(metrics_dict["vllm:request_generation_tokens_sum"], 48)
146-
self.assertEqual(metrics_dict["vllm:request_generation_tokens_bucket"], 3)
161+
self.assertEqual(
162+
metrics_dict["vllm:request_generation_tokens_bucket"], total_prompts
163+
)
147164
# vllm:request_params_best_of
148-
self.assertEqual(metrics_dict["vllm:request_params_best_of_count"], 3)
165+
self.assertEqual(
166+
metrics_dict["vllm:request_params_best_of_count"], total_prompts
167+
)
149168
self.assertEqual(metrics_dict["vllm:request_params_best_of_sum"], 3)
150-
self.assertEqual(metrics_dict["vllm:request_params_best_of_bucket"], 3)
169+
self.assertEqual(
170+
metrics_dict["vllm:request_params_best_of_bucket"], total_prompts
171+
)
151172
# vllm:request_params_n
152-
self.assertEqual(metrics_dict["vllm:request_params_n_count"], 3)
173+
self.assertEqual(metrics_dict["vllm:request_params_n_count"], total_prompts)
153174
self.assertEqual(metrics_dict["vllm:request_params_n_sum"], 3)
154-
self.assertEqual(metrics_dict["vllm:request_params_n_bucket"], 3)
175+
self.assertEqual(metrics_dict["vllm:request_params_n_bucket"], total_prompts)
155176

156177
def test_vllm_metrics_disabled(self):
157178
# Test vLLM metrics

0 commit comments

Comments
 (0)