-
Notifications
You must be signed in to change notification settings - Fork 27.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Any plans to integrate GTE model natively into transformers #35568
Comments
@ArthurZucker @Rocketknight1 If we do intend to integrate this model then I can work on creating a draft PR. |
Is there a place, I can help you to add the model @yaswanth19 ? |
@ArthurZucker A gentle ping. |
This seems popular enough to justify an integration, yes. WDYT @tomaarsen? |
@Rocketknight1
Beyond that, the authors are now using another implementation on top of Qwen:
Some of the mechanisms are similar to ModernBERT (I see unpadding), but some differ as well (xformers). It might require a good bit of effort to get everything to line up with
|
@Rocketknight1 Should I start implementing support for this model, or do you think the effort outweighs the potential benefit and keep using these models with |
Hi @yaswanth19, given @tomaarsen's comment above, I think it's okay to leave them as |
On this topic, the Alibaba team actually just released superior modes based on the new ModernBERT architecture today:
I imagine that they might not move forward with their previous architecture, especially considering they mention that the only parameter they changed for these compared to their previous models was the base model.
|
Model description
Any plans to integrate
gte
model natively into transformers as right now we are using this model withtrust_remote_code=True
argumentOpen source status
Provide useful links for the implementation
Model Implementation: https://huggingface.co/Alibaba-NLP/new-impl/blob/main/modeling.py
Model Weights: https://huggingface.co/Alibaba-NLP/gte-base-en-v1.5
The text was updated successfully, but these errors were encountered: