-
Notifications
You must be signed in to change notification settings - Fork 105
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Model multilingual-e5-small fails to start #583
Comments
@srikanthmanvi @pquentin @technige hey ya'll, wanted to flag that this issue in eland is preventing users from using the e5 small model (easiest to use) in Elasticsearch. It would be amazing if the next version that released for Eland fixed this issue so we can provide support for the e5 small model. |
@serenachou thanks for the ping. We will prioritize this. |
Please note that this applies only to the multilingual E5 model. The normal |
@srikanthmanvi if this isn't on your radar for 8.12 - we would love this to be included into 8.12, or any earlier version that you and @pquentin are cooking up because after 8.12, we would likely be looking to prepared models, so this work would be less effective as a way to encourage customers to use this model for multilingual use cases |
Discussed f2f, @davidkyle will have a look and I will support. |
Any reason to support |
Taking a look at the code updates, it seems this fix should affect the |
@ialdencoots thanks for reporting the problem I have reproduced it myself. You can track the issue at elastic/elasticsearch#102541 The bug fix linked above only applies to the
|
At present, the model is processed and uploaded fine, but when starting the model, it fails:
The model is trained from Multilingual-MiniLM which is a BERT model, but uses the XLM-RoBERTa tokenizer. Since we wrap models based on their architecture and not on the tokenizer type, the BERT model expects input that isn't coming from the XLM-RoBERTa tokenizer. We should consider changing how we decide which wrapper to use (three inputs or two) based on the tokenizer instead.
Note that the base and large model variants work fine because they are XLM-RoBERTa models, and use the corresponding tokenizer.
The text was updated successfully, but these errors were encountered: