You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
If I don't clean up old ASSET_VERSION directories in my bucket, eventually I get to a point where cdn_fu appears to get into an endless loop and never uploads anything. I tracked this to the method:
If the objects.size == MAX_KEYS it enters the loop. When this happens, the method bucket.objects always returns the same set. I've verified this by printing the result on each iteration.
On another (but related) note, is there a reason the uploader has to scan all the existing files even if they are under a different root path (ASSET_VERSION)? As the number of old versions increases, deploys take longer and longer.
using rails 2.3.5 with aws-s3 --version=0.6.2
The text was updated successfully, but these errors were encountered:
If I don't clean up old ASSET_VERSION directories in my bucket, eventually I get to a point where cdn_fu appears to get into an endless loop and never uploads anything. I tracked this to the method:
CloudfrontUploader#populate_existing_asset_checksums
If the objects.size == MAX_KEYS it enters the loop. When this happens, the method bucket.objects always returns the same set. I've verified this by printing the result on each iteration.
On another (but related) note, is there a reason the uploader has to scan all the existing files even if they are under a different root path (ASSET_VERSION)? As the number of old versions increases, deploys take longer and longer.
using rails 2.3.5 with aws-s3 --version=0.6.2
The text was updated successfully, but these errors were encountered: