I am currently setting up a small instance of weaviate to make a proof of concept for myself. I am running my own embedding model for the inserted objects. Running near_vector() works flawlessly, near_text() on the other hand spits out
Query call with protocol GRPC search failed with message panic occurred: ValidateParam was called without any known params present.
I have done some scouring and found that it is possibly related to the fact that I do not have a vectorizor module running on my instance. I know I could recreate my collection with the vectorizer configured but is there a way to specify usage of my own vectorizor?
Continue vectorizing the data yourself, and instead of using near_text, you could vectorize your queries, and use near_vector for search.
Note, you can hybrid search (to run a keyword and vector queries together), but you need to provide the vector embedding for each query.
Also, you can use filters, with near_vector and hybrid.
I was going to do a custom module for Weaviate to have llama.cpp integration but decided I wanted to try to make my own vectorizer work with ollama…and it did! I downloaded a version of my model that was nice and packaged up in the .gguf. I then created a Modelfile the model from said file and the downloaded .gguf.
It now works exactly the same as if I were to use one pulled from Ollama’s repo/library!
I am going to whip up a quick blog post about what I did for the fun of it!