-
Notifications
You must be signed in to change notification settings - Fork 3
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Build a Corpus Inventory #100
Comments
It seems reasonable, although I wonder if we are starting to go down the path of reinventing a wheel that already exists. Would the addition of an indexing solution, such as Elastic Search be another approach? |
It's not about reinventing the wheel. Even though there is solutions like ES and Solr, it's still more efficient to load all informations from one file rather than 1000. :) |
Ok, Just wanted to be sure we weren't overlooking something. |
In order to speed up parsing and reduce parsing time, it might be interesting to build a Generic file merging all data at build. That would reduce the loading time by reducing multiple metadata access.
I can see also a point to group them by a maximum of X (Say 1.000 ? ) :
So, from there, the nautilus resolver, if it detects such file, could default to parsing these insted of using
glob
. It would be a production trick let say...What do you think @balmas @sonofmun
The text was updated successfully, but these errors were encountered: