-
-
Notifications
You must be signed in to change notification settings - Fork 6.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Model] Add SupportsMultiModal.get_language_model
interface
#16007
Conversation
👋 Hi! Thank you for contributing to the vLLM project. 💬 Join our developer Slack at https://slack.vllm.ai to discuss your PR in #pr-reviews, coordinate on features in #feat- channels, or join special interest groups in #sig- channels. Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging. To run CI, PR reviewers can either: Add 🚀 |
This seems like the kind of thing that we could fix on the HF side. Similar to the getter and setter functions we have for input and output embeddings |
How about getting all modalities modules using get_mm_mapping? |
Isn't this entirely dependent on how the model is implemented in vllm?
Unfortunately it's not implemented for every model. But if it were implemented, that would work. I was mostly taking @DarkLight1337 advice. |
You're right, since we reimplement the modelling code in vLLM, having nice utilities in Transformers modelling code doesnt help us here. However, I do see a future where the Transformers backend is stable and performant enough that much of the modelling code in vLLM will not be needed anymore 🤞 |
Yeah I'd be very happy with that future, modelling would be easier. |
We do hope to make model contributions to Transformers easier in future, but yes there may still be some models which need to be modelled in vLLM, which is fine. |
The other PR has been merged, can you update this one? Also there have been a couple new multi-modal models so you should update them as well. |
Signed-off-by: NickLucche <[email protected]>
Signed-off-by: NickLucche <[email protected]>
Signed-off-by: NickLucche <[email protected]>
c3bc449
to
f8eea45
Compare
Rebased and added llama4. I am counting 30 architectures, does that check out? |
You're also missing |
Signed-off-by: NickLucche <[email protected]>
Thanks for looking into it! |
Signed-off-by: NickLucche <[email protected]>
Signed-off-by: NickLucche <[email protected]>
Signed-off-by: NickLucche <[email protected]>
Signed-off-by: DarkLight1337 <[email protected]>
…roject#16007) Signed-off-by: NickLucche <[email protected]> Signed-off-by: zRzRzRzRzRzRzR <[email protected]>
Most vlms adhere to HF unwritten standard and use
self.language_model
, but the naming is not enforced.This PR adds a getter to abstract that naming. See discussion in #15782 (comment) for more context.
I think Whisper is the only outlier in this taxonomy.