-
Notifications
You must be signed in to change notification settings - Fork 22
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Unable to query ai-embed-qa-4
; uninformative error
#30
Comments
this is a good suggestion, also mentioned in #26 for chat completion models |
@aishwaryap is it possible one of the documents you are sending to FAISS is empty? the service rejects empty content. |
@mattf I'm reasonably sure they are not. I took a working example with the That said I can create a self complete example and add it for testing. |
Sample self contained script:
My output (stderr + stdout):
Also verified using
|
@aishwaryap thank you for the reproducer, it helped me narrow this down. i believe the issue is some of the inputs are longer than the embedding model allows. in this case you can pass this is not an issue w/ the does that resolve your issue? |
Hi @mattf, I just found this thread and I just wanted to say that your suggestion worked... at least for the issue at
I get the following error: |
@apolo74 please open this as a new issue, it appears unrelated to embedding and has an informative error |
Hi again @mattf, a couple of minutes ago solved this... I was using a small model (microsoft/phi-3-mini-4k-instruct). There were no more errors the moment I switch to larger models. So the error was related to the size of the LLM. |
@aishwaryap recent changes server-side should have fully resolved this. please reopen this if you still have an issue. |
I am trying to experiment with different embedding models in a RAG application building off of the example here. It works fine when I create an
NVIDIAEmbeddings
object withmodel="nvolveqa_40k"
but withmodel="ai-embed-qa-4"
it fails at the vectorestore creation step iewith the following uninformative error:
I had noticed that for generation models this error sometimes simply means that a newer package version is required and I have filed an issue requesting for more informative errors in that case but with this model, I get this error even with the latest version (0.0.9) and a newly generated API key.
If the model is not yet supported, can it be hidden from the output of
available_models
?The text was updated successfully, but these errors were encountered: