New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
corosync crashes #35
Comments
@michael-dev Do you have reproducer for this bug? Can you please share your config and take a look to log file for error messages? |
Yes this bug happens quite frequently on one virtual machine. This is three corosync nodes (all the same version), two of them run pacemaker. The one failing is the one without pacemaker. All three nodes are connected in two vlans.
I don't have any errors or warnings regarding corosync in syslog. |
@michael-dev Can you please give a try to RRP passive mode? Active rrp is not very well tested. Also passive is better because it makes progress during failure. |
I've changed the cluster configuration and am waiting to see if the bug went away. |
Ok, i changed rpp_mode to passive and otherwise left the config file unchanged.
gdb says:
So the same assert is still hit. |
Any reason for closing this? Has this bug been fixed? |
@michael-dev: Whups. Sorry, I was cleaning old issues where reporter just didn't responded and closed this by mistake. Basically, it's very weird that you are getting so much "Automatically recovered ring 0" so often. Also assert should really never happen. So my theory is, that ether other corosync (probably flatiron) is running on same subnet or packet is corrupted. Can you please try to change mcast port to some different value or you can change mcast address. |
@michael-dev: Were you able to solve this issue or still happening? |
I've not seen this for a while. |
corosync 2.3.3 with libqb-0.17.0 crashes periodically in exec/totemsrp.c:3016, that is
assert (instance->commit_token->memb_index <= instance->commit_token->addr_entries);
backtrace:
The text was updated successfully, but these errors were encountered: