Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

protocol/client: Initialize list head to prevent NULL de-reference #2456

Merged
merged 1 commit into from May 25, 2021

Conversation

pranithk
Copy link
Member

fixes: #2443
Change-Id: I86ef0270d41d6fb924db97fde3196d7c98c8b564
Signed-off-by: Pranith Kumar K pranith.karampuri@phonepe.com

@pranithk pranithk changed the title protocol/client: Initialize list heal to prevent NULL de-reference protocol/client: Initialize list head to prevent NULL de-reference May 20, 2021
@pranithk
Copy link
Member Author

I will try to get an automated test also to prevent this in future.

@pranithk
Copy link
Member Author

/run regression

xhernandez
xhernandez previously approved these changes May 20, 2021
@xhernandez
Copy link
Contributor

@pranithk do you want to provide the test script in this PR ? if not, feel free to merge it.

@pranithk
Copy link
Member Author

@pranithk do you want to provide the test script in this PR ? if not, feel free to merge it.

Yeah, sent the patch in a hurry in the morning based on code-reading. It would be better to wait for the user to let us know the results. I will get some time tomorrow to work on the test script. Will send it by end of tomorrow.

@xhernandez xhernandez dismissed their stale review May 20, 2021 08:45

Removing the approval until the script is added to prevent an accidental merge

@pranithk
Copy link
Member Author

/run regression

@gluster-ant
Copy link
Collaborator

1 test(s) failed
./tests/line-coverage/cli-negative-case-and-function-coverage.t

0 test(s) generated core

5 test(s) needed retry
./tests/000-flaky/basic_ec_ec-quorum-count-partial-failure.t
./tests/000-flaky/glusterd-restart-shd-mux.t
./tests/bugs/glusterd/rebalance-operations-in-single-node.t
./tests/bugs/replicate/bug-880898.t
./tests/line-coverage/cli-negative-case-and-function-coverage.t

1 flaky test(s) marked as success even though they failed
./tests/000-flaky/glusterd-restart-shd-mux.t
https://build.gluster.org/job/gh_centos7-regression/1277/

@pranithk
Copy link
Member Author

Sent #2465 to fix the regression failure. Will wait for that to be merged before running regression run again cc @nik-redhat

fixes: gluster#2443
Change-Id: I86ef0270d41d6fb924db97fde3196d7c98c8b564
Signed-off-by: Pranith Kumar K <pranith.karampuri@phonepe.com>
@pranithk
Copy link
Member Author

/run regression

@gluster-ant
Copy link
Collaborator

1 test(s) failed
./tests/00-geo-rep/01-georep-glusterd-tests.t

0 test(s) generated core

1 test(s) needed retry
./tests/00-geo-rep/01-georep-glusterd-tests.t
https://build.gluster.org/job/gh_centos7-regression/1282/

@pranithk
Copy link
Member Author

/run regression

@pranithk
Copy link
Member Author

@Shwetha-Acharya Could you check ./tests/00-geo-rep/01-georep-glusterd-tests.t is spurious failure?

@Shwetha-Acharya
Copy link
Contributor

@pranithk ./tests/00-geo-rep/01-georep-glusterd-tests.t was failing at EXPECT_WITHIN $GEO_REP_TIMEOUT 1 verify_checkpoint_met primary 127.0.0.1::secondary

This is basically checked from the cli output of geo-replication status in the test case.

I guess the file sync from primary to secondary was slower than the EXPECT_WITHIN time, only in the above scenario. It is not repeatedly observed.

I tried to reproduce this scenario, it was not reproducible anyway on the softserve machines.

@pranithk
Copy link
Member Author

@pranithk ./tests/00-geo-rep/01-georep-glusterd-tests.t was failing at EXPECT_WITHIN $GEO_REP_TIMEOUT 1 verify_checkpoint_met primary 127.0.0.1::secondary

This is basically checked from the cli output of geo-replication status in the test case.

I guess the file sync from primary to secondary was slower than the EXPECT_WITHIN time, only in the above scenario. It is not repeatedly observed.

I tried to reproduce this scenario, it was not reproducible anyway on the softserve machines.

@Shwetha-Acharya Do you want to increase the timeout then?

@pranithk
Copy link
Member Author

@amarts @xhernandez If the changes look okay, could you merge this? I think this should be fixed even in 8.x

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Core dumps on Gluster 9 - 3 replicas
4 participants