Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ensure at least one osd is up #3433

Merged
merged 3 commits into from Feb 19, 2019
Merged

Conversation

dwaiting
Copy link
Contributor

The existing task checks that the number of OSDs is equal to the number of up OSDs before continuing.

The problem is that if none of the OSDs have been discovered yet, the task will exit immediately and subsequent pool creation will fail (num_osds = 0, num_up_osds = 0).

This is related to Bugzilla 1578086.

In this change, we also check that at least one OSD is present. In our testing, this results in the task correctly waiting for all OSDs to come up before continuing.

@gfidente
Copy link
Contributor

check-rdo

1 similar comment
@gfidente
Copy link
Contributor

check-rdo

@softwarefactory-project-zuul
Copy link

Build failed (check pipeline). For information on how to proceed, see
http://docs.openstack.org/infra/manual/developers.html#automated-testing

Copy link
Member

@leseb leseb left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Please add a sign-off to your commit.

The existing task checks that the number of OSDs is equal to the number of up OSDs before continuing.

The problem is that if none of the OSDs have been discovered yet, the task will exit immediately and subsequent pool creation will fail (num_osds = 0, num_up_osds = 0).

This is related to Bugzilla 1578086.

In this change, we also check that at least one OSD is present. In our testing, this results in the task correctly waiting for all OSDs to come up before continuing.

Signed-off-by: David Waiting <david_waiting@comcast.com>
@dwaiting
Copy link
Contributor Author

Still seeing this problem in 3.2.0 and it's causing our OpenStack builds to fail. Can we accept this PR or get some other fix?

@guits
Copy link
Collaborator

guits commented Feb 19, 2019

@dwaiting sorry for the delay. I'll take a look at this in high priority

@mergify mergify bot merged commit 3930791 into ceph:master Feb 19, 2019
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet

4 participants