-
Notifications
You must be signed in to change notification settings - Fork 14
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Refactor search sync to try to reduce memory usage and hash ids #662
Conversation
57d2ecb
to
7371b4c
Compare
7371b4c
to
8b84adb
Compare
There's a lot of code in |
Yeah
Of course, all attributes that can be searched on, or are returned, must be stored in Meilisearch. It not as simple as doing a union of 2. and 3. to get 1. however, since some attributes (like Then there are objects like I know that the file is full of types, but it is to prevent us developers from accidentally trying to e.g. do custom ranking rules on an attribute that isn't even stored in Meilisearch. As for |
Thanks! I added that as a comment to the top of the file since it helped a lot. |
Title speaks for itself.
Idea is to divide data in batches and send it one at a time, this way the garbage collector should be able to drop objects since they go out of scope.
Current batch size is 1000, which I think should be fine.
Also hashes ids which should be a (temporary) fix to #661