Skip to content

Commit 5608e61

Browse files
authored
[Doc] Update Qwen documentation (#9869)
1 parent 3ea2dc2 commit 5608e61

File tree

2 files changed

+6
-3
lines changed

2 files changed

+6
-3
lines changed

docs/source/models/supported_models.rst

Lines changed: 5 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -277,7 +277,7 @@ Text Generation
277277
* - :code:`QWenLMHeadModel`
278278
- Qwen
279279
- :code:`Qwen/Qwen-7B`, :code:`Qwen/Qwen-7B-Chat`, etc.
280-
-
280+
- ✅︎
281281
- ✅︎
282282
* - :code:`Qwen2ForCausalLM`
283283
- Qwen2
@@ -516,7 +516,7 @@ Text Generation
516516
- Qwen-VL
517517
- T + I\ :sup:`E+`
518518
- :code:`Qwen/Qwen-VL`, :code:`Qwen/Qwen-VL-Chat`, etc.
519-
-
519+
- ✅︎
520520
- ✅︎
521521
* - :code:`Qwen2AudioForConditionalGeneration`
522522
- Qwen2-Audio
@@ -540,6 +540,9 @@ Text Generation
540540
| :sup:`E` Pre-computed embeddings can be inputted for this modality.
541541
| :sup:`+` Multiple items can be inputted per text prompt for this modality.
542542
543+
.. note::
544+
vLLM currently only supports adding LoRA to the language backbone of multimodal models.
545+
543546
.. note::
544547
For :code:`openbmb/MiniCPM-V-2`, the official repo doesn't work yet, so we need to use a fork (:code:`HwwwH/MiniCPM-V-2`) for now.
545548
For more details, please see: https://github.com/vllm-project/vllm/pull/4087#issuecomment-2250397630

vllm/model_executor/models/qwen.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1048,7 +1048,7 @@ def get_mm_mapping(self) -> MultiModelKeys:
10481048
@MULTIMODAL_REGISTRY.register_max_image_tokens(MAX_QWEN_IMG_TOKENS)
10491049
@INPUT_REGISTRY.register_dummy_data(dummy_data_for_qwen)
10501050
@INPUT_REGISTRY.register_input_processor(input_processor_for_qwen)
1051-
class QWenLMHeadModel(QWenBaseModel):
1051+
class QWenLMHeadModel(QWenBaseModel, SupportsLoRA):
10521052
"""
10531053
QWenLMHeadModel is not only applicable to LLM but also to VL, which is not
10541054
conducive to the current integration logic of LoRA in vLLM. Therefore, it

0 commit comments

Comments
 (0)