-
Notifications
You must be signed in to change notification settings - Fork 13.1k
Support conversion of Qwen3-Embedding models #15023
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for looking into this - it would be great to get this model supported.
Ooops, was a bit trigger happy, forgot @iamlemec care to add the model there in a new PR? Edit: It has to go in |
@iamlemec Did you ever speak to Qwen about the official GGUFs? Are they okay to use in their current form or would they have to be updated? |
@Mushoz looks like they uploaded new GGUFs to HF! Just tested the 0.6B version and it works as expected with correct pooling. |
@iamlemec which ones are you talking about? I am looking at this one: https://huggingface.co/Qwen/Qwen3-Embedding-0.6B-GGUF/tree/main Same holds true for the 8b version: https://huggingface.co/Qwen/Qwen3-Embedding-8B-GGUF/tree/main |
@Mushoz Huh, you're right. Maybe I missed that update before and was still using the old ones? Anyway, the ones you linked to have the pooling type properly specified and seem to work as expected! |
This adds GGUF conversion support for the Qwen3-Embedding class of models and makes pooling work properly by default. I'm not sure how the official Qwen GGUFs were produced, but I'll try to get them updated if this gets merged. More details:
qwen2
as usual, the HF tokenizer adds an EOT token that makes the checksum different from the usual text generation modelsmodel.
like most other models in this class