-
-
Notifications
You must be signed in to change notification settings - Fork 901
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
MD5 cache file not updated after each calculated hash #641
Comments
Yes, the cache is stored as a python pickle, at the end of the local On Fri, Oct 2, 2015 at 5:47 PM, Ramon Smits notifications@github.com
|
Not familiar with python but is a pickle sort of a hash table or Is there a graceful abort possible that will still write the pickle to disk A solution would be a log file that gets appended. At start you load the After the file scan, store the pickle, delete the log file. Another option would be to store the pickle for a given interval so that if I prefer the first. It explains why people mention that the md5 cache is not working. I'm now Let me dive into python, maybe I can contribute to s3cmd.
|
Today my wifi connection failed and the sync quit. No md5 hashes were flushed to disk at all. Also, having a binary file that is written to at the end of the batch makes it impossible for multiple invocations to share the same cache file. An alternative is to create a .md5 file for each file and
Or use a file per folder and maybe even use the same file format as md5sum as text. Or use a file per tree |
wow this does not seem to have gotten any love in a long time. was there ever a resolution? I am currently facing the same issue. every time I run the sync command it has to calculate 40k md5 hashes... |
Same issue here. When syncing 8000 files s3cmd seems to generate an md5 every time and doesn't save the MD5 results to cache on local disk (s3cmd version 2.3.0). It makes for really long sync times. Does anyone have any ideas? |
I'm currently running a
s3cmd sync
operation on my laptop from s3 to my nas. The operation is running for hours because the md5 hashes need to be generated. I added the--cache-file
option so that the md5 hashes will be stored.However, I just looked at the folder and I don't see that file. Does that mean that the cache file is only flushed to storage at the end of the sync operation?
Why isn't the cache file flushed after each calculated md5? I cannot stop the current operation as then all calculated md5 hashes are gone.
I'm running the operation with
--verbose
and after almost 11 hours it says 9000/40244 so that means it is now at 22%.The text was updated successfully, but these errors were encountered: