New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Index shard got corrupted #186
Comments
Guys any update ? |
Which version of elasticsearch was the snapshot created with? |
We are using On Thu, Mar 5, 2015 at 2:40 AM, Igor Motov notifications@github.com wrote:
|
@sukantasaha I understand that 1.4.1 is the version that you are currently using. But I am trying to figure out which version of elasticsearch this index was first snapshotted with. You said that you are snapshotting on a regular basis. So, judging from the name, the index was created in June 2014. Is this when you created the first snapshot of this index or was it later on? If this index was first snapshotted back then and you were staying with the latest (more or less) available version of elasticsearch we can assume that this index was first snapshotted by elasticsearch v1.1 or v1.2. Was this the case? |
This cluster was created on Thu, Jul 2014 with elasticsearch-1.2.1 and may On Thu, Mar 5, 2015 at 10:33 AM, Igor Motov notifications@github.com
|
@sukantasaha what did you mean by "may be we migrated this index"? Could it be possible that this index was originally created with older version of elasticsearch? I am just wondering if you are hitting elastic/elasticsearch#9140. Would it be possible for you as an experiment to restore this index using a newer version of elasticsearch (v1.4.3 or higher)? |
We did the same test, Launched a new cluster with 1.4.3 version and there If you want I can do the same again and show you On Sat, Mar 7, 2015 at 8:26 AM, Igor Motov notifications@github.com wrote:
|
May I know that when S3 plugin has been extended by MD5 checksum and in which version , so that no more problems with snapshot restore because we get to know if snapshot was really successful or not |
MD5 could be added in 2.4.2, so it's not there yet. |
When can we expect ? |
No ETA yet. And the PR has not been merged. |
Is this fix now in es-1.5? |
No. It's not. |
There is a feature available in S3 that clients can use to ensure data integrity on upload. Whenever an object is PUT to an S3 bucket, the client is able to get back the `MD5` base64 encoded and check that it's the same `MD5` as the local one. For reference, please see the [S3 PutObject API](http://docs.aws.amazon.com/AmazonS3/latest/API/RESTObjectPUT.html). Closes #186. (cherry picked from commit 2d4fd39) (cherry picked from commit 3369e02) (cherry picked from commit 1d1f5c7)
There is a feature available in S3 that clients can use to ensure data integrity on upload. Whenever an object is PUT to an S3 bucket, the client is able to get back the `MD5` base64 encoded and check that it's the same `MD5` as the local one. For reference, please see the [S3 PutObject API](http://docs.aws.amazon.com/AmazonS3/latest/API/RESTObjectPUT.html). Closes #186. (cherry picked from commit 2d4fd39)
There is a feature available in S3 that clients can use to ensure data integrity on upload. Whenever an object is PUT to an S3 bucket, the client is able to get back the `MD5` base64 encoded and check that it's the same `MD5` as the local one. For reference, please see the [S3 PutObject API](http://docs.aws.amazon.com/AmazonS3/latest/API/RESTObjectPUT.html). Closes #186. (cherry picked from commit 2d4fd39) (cherry picked from commit 3369e02)
Hi
in all our elasticsearch cluster we use this elasticsearch-cloud-aws plugin to create the snapshots on s3 on a regular basis.
Some times we saw the shard got corrupted for an index in our elasticsearch log.
So we try to restore it from backup and while restoring it from backup again we see the same exception in logs which is follows
Even if we back to an older snapshot we found the same exception.
So what we did was we download all the segments files from s3 merge it and there we found some segments were corrupted by using
org.apache.lucene.index.CheckIndex
with-fix
We fixed it but we loose 5gb data.
We shared this problem with aws team, here what they suggested:
Can you guys please have a look into this issue and suggest something
Thanks
The text was updated successfully, but these errors were encountered: