Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add support for 'rotary' position embedding type for xml-roberta models #3536

Open
Aelentel opened this issue Nov 17, 2024 · 0 comments
Open
Labels
enhancement New feature or request

Comments

@Aelentel
Copy link

Description

After converting the hugging face model "jinaai/jina-embeddings-v3" via djl-convert and patching the missing model type to "xml-roberta" it still barf on the unknown position embedding type.

  • Rust conversion didn't crash but didn't yield an usable model.
  • PyTorch crashed
  • OnnxRuntime crashed

version of djl-convert tested : 0.30.0 and 0.31.0

Will this change the current api? : absolutely no idea.

Who will benefit from this enhancement?
any person using xml-roberta with "rotary" position embedding type

also this model is VERY efficient at embedding multilanguages and cross language semantic, i tested it on 20k docs on 13 different languages and the semantic matching was awesome, it'll be very good to support it.

References

  • works using the following code on python langchain.
from langchain_huggingface import HuggingFaceEmbeddings

model_name = "jinaai/jina-embeddings-v3"
model_kwargs = {"device": "cuda","trust_remote_code":True}

hf = HuggingFaceEmbeddings(model_name=model_name, model_kwargs=model_kwargs)

i'm pretty new on AI models so feel free to fix any of my mistakes :)

@Aelentel Aelentel added the enhancement New feature or request label Nov 17, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant