-
Notifications
You must be signed in to change notification settings - Fork 10.8k
S3 Feed Export throws boto error: "Connection Reset By Peer" #960
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Scrapy uses boto for feed exports, so it is likely a boto issue (the one you've linked at SO, boto/boto#2207). Do you know a workaround? |
To send big feed files to S3 and avoid this bug, the fix is to use multipart upload. PR #1559 used to implement this for boto2. |
It would be great if someone can resurrect this WIP #1559 |
Looking at the upstream API, it seems like implementing this change is a matter of:
|
#4077 is an interesting approach, which starts uploading to S3 right away, rather than storing the whole output on disk first, and then uploading it. But I would rather have a simpler approach than none. We can always implement the #4077 approach afterwards. |
See also the workaround by @ogabrielsantos. |
I'd like to have a go at the simpler approach, if that's alright. |
@Gallaecio Hey, I did the "check libraries availability" part, it passes the tests with |
It sounds like you are. But feel free to open a draft pull request, and I can have a look. |
Fix #960: S3 Feed Export throws boto error
Posted details on SO: http://stackoverflow.com/questions/27131693/scrapyd-s3-feed-export-connection-reset-by-peer
The text was updated successfully, but these errors were encountered: