Skip to content

Commit f211331

Browse files
reidliu41reidliu41
and
reidliu41
authored
[Doc] small fix (#17277)
Signed-off-by: reidliu41 <[email protected]> Co-authored-by: reidliu41 <[email protected]>
1 parent 9053d0b commit f211331

File tree

2 files changed

+3
-1
lines changed

2 files changed

+3
-1
lines changed

docs/source/models/generative_models.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -59,7 +59,7 @@ A code example can be found here: <gh-file:examples/offline_inference/basic/basi
5959

6060
### `LLM.beam_search`
6161

62-
The {class}`~vllm.LLM.beam_search` method implements [beam search](https://huggingface.co/docs/transformers/en/generation_strategies#beam-search-decoding) on top of {class}`~vllm.LLM.generate`.
62+
The {class}`~vllm.LLM.beam_search` method implements [beam search](https://huggingface.co/docs/transformers/en/generation_strategies#beam-search) on top of {class}`~vllm.LLM.generate`.
6363
For example, to search using 5 beams and output at most 50 tokens:
6464

6565
```python

docs/source/models/supported_models.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -793,6 +793,8 @@ or `--limit-mm-per-prompt` (online serving). For example, to enable passing up t
793793
Offline inference:
794794

795795
```python
796+
from vllm import LLM
797+
796798
llm = LLM(
797799
model="Qwen/Qwen2-VL-7B-Instruct",
798800
limit_mm_per_prompt={"image": 4},

0 commit comments

Comments
 (0)