Replies: 1 comment 1 reply
-
One of my colleagues just told me about Nvidia Triton which is thought for handling different models at the same time. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Idea
Many people probably have the same needs, that is:
I was wondering about your thoughts considering some kind of "smooth" integration of e.g. sentence-transformers (GitHub here). The same could obviously also be done for image encodings.
It would be charming if Qdrant somehow could integrate these model directly in its API, so internalizing step 2 also in Qdrant.
E.g. when initializing a new collection, one could optionally add any public transformer model and use it as a default in the API.
Advantages
Instead of having e.g. a Python microservice somewhere (or in whatever language the wrapper is defined) that:
this could be done all from the Qdrant API and would very likely be faster than the above "standard" way.
It would be so convenient if you think about it!
Writing to DB
When writing to the DB it would be very nice to simply send an array of documents and Qdrant takes care of all the rest.
Reading from DB
Reading is the biggest advantage from my point of view. You could create a fully-fledged app with just Qdrant and a static frontend!
In the frontend, a user would only input a search term and fire a request to the API.
Qdrant would calculate the embedding for the search term, search the DB and respond with the results, all in one go!
Challenges
I am aware that it might be way out of scope for Qdrant but I think it also has the potential to attract a very large user base if it's that beginner-friendly.
A challenge is probably that a model must first be downloaded, initialized and running somewhere which might be tricky to set up in one Docker image.
A more feasible solution would maybe be some kind of link between the Qdrant Docker image and another Docker image serving the model that could be set up in advance if needed.
What do you think about this?
Beta Was this translation helpful? Give feedback.
All reactions