Skip to content

Handling huge embeddings databases. #261

Answered by vladmandic
lambdacerro asked this question in Q&A
Discussion options

You must be logged in to vote

I was thinking to use postgresql since it does have native support for array but 1024 dimensions is too much for it to handle, another options that comes to mind is MongoDB.

I've used MongoDB in several projects, its pretty trivial and quite fast.

Most of the other facereg libraries I've used deals with 128 dimensions for their embedding however human uses 1024 arrays, this does complicate things a little bit.

You could reduce number of computed dimensions. Matching algorithm cares that both source and target descriptor have same number of dimensions, but they can be anything.

Reducing dimensions does decrease precision, but you can play with what is acceptable to you (1024 -> 512 -> …

Replies: 3 comments

Comment options

You must be logged in to vote
0 replies
Answer selected by lambdacerro
Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants