New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Segmentation fault on shard 2 during node decommission #5089
Comments
@eliransin please do the triage |
@yarongilor is this segfault the same one hit in #5087? |
@amoskong / @avikivity could toppartitions be involved here too? (#5104) |
@bhalevy it is not the same as #5087 |
The suspicious logic is: |
Duplicate of #4685? @vladzcloudius? |
If I am not mistaken Vlad sent the fix for this almost 2 months ago.
It is still waiting for reviews.
…On Mon, Sep 30, 2019 at 6:08 AM Piotr Sarna ***@***.***> wrote:
Duplicate of #4685? @vladzcloudius?
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub, or mute the thread.
|
@glommer true, I only found it today and reviewed. |
@bhalevy yes #5087 points to the same time in log - 2019-09-20T01:34 - 37 |
patch merged d9dc8f9 and backported - closing issue will be part of 3.1.rc9 |
Installation details
Scylla version (or git commit hash): 3.1.0.rc7-0.20190915.024d1563a
Cluster size: 6
OS (RHEL/CentOS/Ubuntu/AWS AMI): ami-0a01b94dffe4bcc0d
scenario: 50GB longevity - during nodetool_decommission nemesis of 10.0.63.76 node, there was a core dump on 10.0.244.153 node. then 10.0.244.153 got "Segmentation fault on shard 2"
failure scenario and error:
segmentation fault backtrace:
decoded backtrace:
coredump details: download_instructions=
gsutil cp gs://upload.scylladb.com/core.scylla.996.c212eda51d7d40c1b68eb746938de6ae.1730.1568943298000000/core.scylla.996.c212eda51d7d40c1b68eb746938de6ae.1730.1568943298000000.gz .
test logs:
job log: scratch.scylladb.com/qa_issues/sct_08_42_22_09_2019.log.gz
snapshot:
https://snapshot.raintank.io/dashboard/snapshot/I6ue7n0F4RQkowyEhHqc2p44NdxIZwT1
The text was updated successfully, but these errors were encountered: