We use Cloudflare Workers for our API and just handle auth calls by checking the JWTs with Supabase and caching it. So we already had the necessary auth setup to do this.
For basic CRUD we use the Supabase endpoints directly but none of that involves querying a vector db :P
> Qdrant stores both the vectors and the metadata in a sqlite database.
It looks like you have benchmarked local mode of qdrant.
It doesn't even use vector indexes and is not designed for any kind of production usage.
For anyone reading this article, I urge you to do your own benchmarks and not rely on claims that do not have open source code attached to them to replicate the results
Hi Andrey. Thanks for your feedback. We should have better emphasized that we are benchmarking Qdrant in local mode. We have updated the post to clarify that Qdrant is being evaluated in local mode. We plan to next evaluate the server mode.
Made it easier to document code. You can send the function as context, and it will generate a well-formatted docstring that accomplishes 90% of the task!
I appreciate your view on using LLMs in production. However, I think the ability to use ML/AI directly within databases is a significant step forward. Most production systems already rely on predictions, so the idea of using them without leaving the database is both interesting and valuable.
It probably could but it would require adjusting the prompt. You’d have to override the generate_prompt function and tell it that you’re using a variant of SQL and describe the differences.
There isn't technically a limit on the storage side but it's generally better if you keep documentation to a manageable length.
You call vn.train(sql=...) on each individual SQL statement that you have.
What'll happen under the hood is the package will use the 10 most relevant SQL statement examples, 10 most relevant pieces of documentation, and 10 most relevant DDL statements.
If using 10 examples exceeds the (approximate) token limit for the model, it'll pare down to a smaller number that'll fit into the context limit.