New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
mxbai-embed-large embedding not consistent with original paper #4207
Comments
Same thing with snowflake: Ollama using
I expected these to give same results. |
I've now also verified that the embeddings generated by https://github.com/ggerganov/llama.cpp/tree/master/examples/embedding are correct and consistent with the blog post: ./embedding --model ./models/mxbai-embed-large/mxbai-embed-large-v1-f16.gguf --prompt $'Represent this sentence for searching relevant passages: A man is eating a piece of bread\nA man is eating food.\nA man is eating pasta.\nThe girl is carrying a baby.\nA man is riding a horse.' Output:
|
I've now also confirmed the issue is still happening in ollama v0.1.34 |
#3777 |
I also have this issue v0.1.35. following the blog with mxbai-embed-large gives the wrong results. switching embedding models provide the correct results |
I can also confirm that Ollama embeddings for
I'm not sure if the result from Edit: My ollama version is 0.1.28 |
Similarities after PR #4399: mxbai-embed-large
snowflake-arctic-embed:137m-m-long-fp16
|
Tack @deadbeef84 for creating this issue. I've been scratching my head over the weekend, changed to different embedding models and had very odd results. Good to know I wasn't crazy and that Ollama actually was a bit broken. I'll update Ollama and see if my ChromaDB <-> Ollama experiment runs better. Update: I downloaded and built #4399. I can confirm it indeed fixes obvious issues I had when doing embedding and queries with |
@deadbeef84 thanks for the fix, this solves the issue. |
What is the issue?
I'm trying to use embeddings from
mxbai-embed-large
to create a similarity/semantic search functionality, but the quality of the embeddings coming from ollama doesn't seem to be very good.I've tried replicating the numbers from the original blog post:
Those numbers are nowhere close to the original numbers, and if I compare the embedding vectors they are completely different.
The javascript implementation at huggingface produces the same numbers as the original post.
OS
Linux, Docker
GPU
Nvidia
CPU
AMD
Ollama version
0.1.33
The text was updated successfully, but these errors were encountered: