We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
gte-Qwen2
1 parent 40c1080 commit 20bd63cCopy full SHA for 20bd63c
docs/source/models/supported_models.md
@@ -430,6 +430,9 @@ You can set `--hf-overrides '{"is_causal": false}'` to change the attention mask
430
431
On the other hand, its 1.5B variant (`Alibaba-NLP/gte-Qwen2-1.5B-instruct`) uses causal attention
432
despite being described otherwise on its model card.
433
+
434
+Regardless of the variant, you need to enable `--trust-remote-code` for the correct tokenizer to be
435
+loaded. See [relevant issue on HF Transformers](https://github.com/huggingface/transformers/issues/34882).
436
```
437
438
If your model is not in the above list, we will try to automatically convert the model using
0 commit comments