Which model is everyone using for embeddings? #33
Replies: 1 comment 1 reply
-
I have had vastly different results between OpenAI Text 3 Large and everything else - my chunking strategy is now so aligned with their current context window and response token limit that I am not even considering investing time in the others. It's a bit expensive but I don't want to risk using a smaller model API for fear of not maximizing my chances. I embed both my notes and blocks with this model and set a lower threshold for embedding of 10. My data is comprised of a lot of standards and specification so a lot of it is pre-chunked with a lot of small chunks that are relevant so warrant their own embedding in a document hierarchy - as I've curated the data I've learned a lot about how to deploy a successful chunking strategy |
Beta Was this translation helpful? Give feedback.
-
Just curious, which models have you guys tested for notes/blocks and have you found any core difference in speed/recall/accuracy?
Beta Was this translation helpful? Give feedback.
All reactions