diff --git a/docs/source/models/supported_models.md b/docs/source/models/supported_models.md index 7145bcf2d5f..b046ccfd155 100644 --- a/docs/source/models/supported_models.md +++ b/docs/source/models/supported_models.md @@ -726,7 +726,7 @@ See [this page](#generative-models) for more information on how to use generativ * `h2oai/h2ovl-mississippi-800m`, `h2oai/h2ovl-mississippi-2b`, etc. * * ✅︎ - * \* + * ✅︎\* - * `Idefics3ForConditionalGeneration` * Idefics3 * T + I @@ -869,7 +869,7 @@ See [this page](#generative-models) for more information on how to use generativ + Multiple items can be inputted per text prompt for this modality. :::{note} -H2O-VL series models will be available in V1 once we support backends other than FlashAttention. +`h2oai/h2ovl-mississippi-2b` will be available in V1 once we support backends other than FlashAttention. ::: :::{note} diff --git a/examples/offline_inference/vision_language.py b/examples/offline_inference/vision_language.py index b9963669a0d..5f05389faf8 100644 --- a/examples/offline_inference/vision_language.py +++ b/examples/offline_inference/vision_language.py @@ -119,7 +119,7 @@ def run_glm4v(question: str, modality: str): def run_h2ovl(question: str, modality: str): assert modality == "image" - model_name = "h2oai/h2ovl-mississippi-2b" + model_name = "h2oai/h2ovl-mississippi-800m" llm = LLM( model=model_name, @@ -136,7 +136,7 @@ def run_h2ovl(question: str, modality: str): add_generation_prompt=True) # Stop tokens for H2OVL-Mississippi - # https://huggingface.co/h2oai/h2ovl-mississippi-2b + # https://huggingface.co/h2oai/h2ovl-mississippi-800m stop_token_ids = [tokenizer.eos_token_id] return llm, prompt, stop_token_ids diff --git a/examples/offline_inference/vision_language_multi_image.py b/examples/offline_inference/vision_language_multi_image.py index 1a5ea0c70bc..b2821966cf1 100644 --- a/examples/offline_inference/vision_language_multi_image.py +++ b/examples/offline_inference/vision_language_multi_image.py @@ -78,14 +78,13 @@ def load_deepseek_vl2(question: str, image_urls: List[str]): def load_h2ovl(question: str, image_urls: List[str]) -> ModelRequestData: - model_name = "h2oai/h2ovl-mississippi-2b" + model_name = "h2oai/h2ovl-mississippi-800m" llm = LLM( model=model_name, trust_remote_code=True, max_model_len=8192, limit_mm_per_prompt={"image": len(image_urls)}, - mm_processor_kwargs={"max_dynamic_patch": 4}, ) placeholders = "\n".join(f"Image-{i}: \n" @@ -99,7 +98,7 @@ def load_h2ovl(question: str, image_urls: List[str]) -> ModelRequestData: add_generation_prompt=True) # Stop tokens for H2OVL-Mississippi - # https://huggingface.co/h2oai/h2ovl-mississippi-2b + # https://huggingface.co/h2oai/h2ovl-mississippi-800m stop_token_ids = [tokenizer.eos_token_id] return ModelRequestData(