Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Backfill database from existing ElasticSearch cluster #36

Open
Raxvis opened this issue Oct 19, 2018 · 19 comments
Open

Backfill database from existing ElasticSearch cluster #36

Raxvis opened this issue Oct 19, 2018 · 19 comments

Comments

@Raxvis
Copy link
Contributor

Raxvis commented Oct 19, 2018

No description provided.

@Raxvis
Copy link
Contributor Author

Raxvis commented Oct 19, 2018

Include votes and flags

@ghost
Copy link

ghost commented Jan 29, 2019

I'm still searching for a way to export elasticsearch data to mysql. Do you know of any way to so that?
Kind regards

@Raxvis
Copy link
Contributor Author

Raxvis commented Jan 29, 2019

I was going to build a script from this, but I have gotten a new computer since then and don't have your backup file anymore. Would you prefer a backfill from an export, or backfill from scrolling through elasticsearch itself?

@ghost
Copy link

ghost commented Jan 29, 2019

No worries :) I don't mind whatever is faster I guess.

@ghost
Copy link

ghost commented Jan 30, 2019

I could provide a smaller backup file from the old scraper if you like ?

@Raxvis
Copy link
Contributor Author

Raxvis commented Jan 30, 2019 via email

@ghost
Copy link

ghost commented Jan 31, 2019

sorry got caught up in other stuff. ill get you a backup tonight.

@Raxvis
Copy link
Contributor Author

Raxvis commented Jan 31, 2019 via email

@ghost
Copy link

ghost commented Feb 1, 2019

Was just thinking if used the backup I have of 20 million torrents and imported them to the new database . then wouldn't i end up with the same issue with tons of torrents not updated with seeders/leechers info.

also just putting an idea out there. what about feeding the scraper a backup file of just infohashes? could this be done? id pay for this :D @Prefinem

@milezzz
Copy link

milezzz commented Feb 1, 2019

what about feeding the scraper a backup file of just infohashes?

That would be really amazing! /me tosses in some BTC.

@Raxvis
Copy link
Contributor Author

Raxvis commented Feb 1, 2019

That should be doable.

If you can get me a backup file, I can create an import script. I lost the last one you sent me when I got a new computer.

So... just to summarize:

  • Restore from elasticsearch backup (one time process)
  • Import list of infohashes (script)

Would that work?

@ghost
Copy link

ghost commented Feb 1, 2019

That would be perfect :) I'll do the database today for you.

And yes if we could feed it just a ton of infohashes and it could run them through and get all the meta info etc?

Send me your PayPal or Bitcoin and I'll donate some

@ghost
Copy link

ghost commented Feb 1, 2019

@Prefinem here is a small database.

torrents.zip

@ghost
Copy link

ghost commented Feb 2, 2019

Is that ok for you mate ?

@Raxvis
Copy link
Contributor Author

Raxvis commented Feb 2, 2019 via email

@ghost
Copy link

ghost commented Feb 2, 2019

The large back up file I have is around 25gb I believe. Would this be too large ? If so maybe it would be better to export to mysql by querying elasticsearch directly?

Kind regards

@Raxvis
Copy link
Contributor Author

Raxvis commented Feb 2, 2019 via email

@ghost
Copy link

ghost commented Feb 2, 2019

Sure I'll send you the ip.

@Raxvis
Copy link
Contributor Author

Raxvis commented Feb 2, 2019 via email

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants