Skip to content

Commit 49ebd23

Browse files
hhy3lk-chen
authored andcommitted
[Bugfix] Fix profiling.py (vllm-project#16202)
Signed-off-by: zh Wang <[email protected]>
1 parent 0e2bd79 commit 49ebd23

File tree

1 file changed

+2
-3
lines changed

1 file changed

+2
-3
lines changed

examples/offline_inference/profiling.py

Lines changed: 2 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -234,9 +234,8 @@ def get_output_len_generator() -> Generator[int, Any, Any]:
234234
sampling_params.max_tokens = next(output_len_generator)
235235
assert isinstance(sampling_params.max_tokens, int)
236236

237-
prompt_token_ids = torch.randint(
238-
llm.llm_engine.model_config.get_vocab_size(),
239-
size=(prompt_len, )).tolist()
237+
prompt_token_ids = torch.randint(llm.get_tokenizer().vocab_size,
238+
size=(prompt_len, )).tolist()
240239

241240
llm.llm_engine.add_request(
242241
request_id=f"seq{i}",

0 commit comments

Comments
 (0)