Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Investigate faster ways to renormalize. #19

Open
johnnykv opened this issue Apr 5, 2013 · 0 comments
Open

Investigate faster ways to renormalize. #19

johnnykv opened this issue Apr 5, 2013 · 0 comments

Comments

@johnnykv
Copy link
Owner

johnnykv commented Apr 5, 2013

At the current state (15.000.000 hpfeed entries) it takes around 7 hours to do a full database renormalization (--reset). A database renormalization consists of:

  • Dropping all collections except for the hpfeed collection
  • Recreating indexes
  • Processing of all hpfeed entries to populate the dork, file, url and session collection.
  • Incrementing alot of counters in the daily_stats collection.

Could this be done faster than 7 hours on the current hardware? (around 2.1 million hpfeed entried pr. hour)

jatrost added a commit to jatrost/mnemosyne that referenced this issue Dec 27, 2016
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant