diff --git a/docs/source/models/supported_models.md b/docs/source/models/supported_models.md
index 7145bcf2d5f..b046ccfd155 100644
--- a/docs/source/models/supported_models.md
+++ b/docs/source/models/supported_models.md
@@ -726,7 +726,7 @@ See [this page](#generative-models) for more information on how to use generativ
* `h2oai/h2ovl-mississippi-800m`, `h2oai/h2ovl-mississippi-2b`, etc.
*
* ✅︎
- * \*
+ * ✅︎\*
- * `Idefics3ForConditionalGeneration`
* Idefics3
* T + I
@@ -869,7 +869,7 @@ See [this page](#generative-models) for more information on how to use generativ
+ Multiple items can be inputted per text prompt for this modality.
:::{note}
-H2O-VL series models will be available in V1 once we support backends other than FlashAttention.
+`h2oai/h2ovl-mississippi-2b` will be available in V1 once we support backends other than FlashAttention.
:::
:::{note}
diff --git a/examples/offline_inference/vision_language.py b/examples/offline_inference/vision_language.py
index b9963669a0d..5f05389faf8 100644
--- a/examples/offline_inference/vision_language.py
+++ b/examples/offline_inference/vision_language.py
@@ -119,7 +119,7 @@ def run_glm4v(question: str, modality: str):
def run_h2ovl(question: str, modality: str):
assert modality == "image"
- model_name = "h2oai/h2ovl-mississippi-2b"
+ model_name = "h2oai/h2ovl-mississippi-800m"
llm = LLM(
model=model_name,
@@ -136,7 +136,7 @@ def run_h2ovl(question: str, modality: str):
add_generation_prompt=True)
# Stop tokens for H2OVL-Mississippi
- # https://huggingface.co/h2oai/h2ovl-mississippi-2b
+ # https://huggingface.co/h2oai/h2ovl-mississippi-800m
stop_token_ids = [tokenizer.eos_token_id]
return llm, prompt, stop_token_ids
diff --git a/examples/offline_inference/vision_language_multi_image.py b/examples/offline_inference/vision_language_multi_image.py
index 1a5ea0c70bc..b2821966cf1 100644
--- a/examples/offline_inference/vision_language_multi_image.py
+++ b/examples/offline_inference/vision_language_multi_image.py
@@ -78,14 +78,13 @@ def load_deepseek_vl2(question: str, image_urls: List[str]):
def load_h2ovl(question: str, image_urls: List[str]) -> ModelRequestData:
- model_name = "h2oai/h2ovl-mississippi-2b"
+ model_name = "h2oai/h2ovl-mississippi-800m"
llm = LLM(
model=model_name,
trust_remote_code=True,
max_model_len=8192,
limit_mm_per_prompt={"image": len(image_urls)},
- mm_processor_kwargs={"max_dynamic_patch": 4},
)
placeholders = "\n".join(f"Image-{i}: \n"
@@ -99,7 +98,7 @@ def load_h2ovl(question: str, image_urls: List[str]) -> ModelRequestData:
add_generation_prompt=True)
# Stop tokens for H2OVL-Mississippi
- # https://huggingface.co/h2oai/h2ovl-mississippi-2b
+ # https://huggingface.co/h2oai/h2ovl-mississippi-800m
stop_token_ids = [tokenizer.eos_token_id]
return ModelRequestData(