New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
rolling_update: migrate ceph-disk osds to ceph-volume #3727
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think you need to add a condition because this will only work for non-containerized deployment.
c7b9c76
to
52c281c
Compare
You're right, I've got that added now. |
dc8eeba
to
93ee82e
Compare
|
The |
30ca2ba
to
c9f9446
Compare
|
I manually ran the new testing sceneario added here and it was successful. https://2.jenkins.ceph.com/job/ceph-ansible-scenario/548/consoleFull You can see the There is a caveat to upgrading to nautilus. For the upgrade to complete, the user must switch the |
c9f9446
to
00ad57f
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The idea in ceph-ansible@master is to test an upgrade from ceph@nautilus to ceph@dev
The same scenario to upgrade ceph@luminous to ceph@nautilus will be tested in ceph-ansible@stable-4.0
What value are we really gaining by running tests that upgrade from nautilus to ceph@master? The point is that I must upgrade from luminous (or mimic) to nautilus so that I can verify that the new |
|
so it should be done in |
I'm not sure I understand. Are you saying that I should deploy luminous with |
00ad57f
to
f7ada05
Compare
|
jenkins test pipeline |
c89812d
to
6ba3345
Compare
When upgrading to nautlius run ``ceph-volume simple scan`` and ``ceph-volume simple activate --all`` to migrate any running ceph-disk osds to ceph-volume. Closes: https://bugzilla.redhat.com/show_bug.cgi?id=1656460 Signed-off-by: Andrew Schoen <aschoen@redhat.com>
This test deploys a luminous cluster with ceph-disk created osds and then upgrades to nautilus and migrates those osds to ceph-volume. The nodes are then rebooted and cluster state verified. Signed-off-by: Andrew Schoen <aschoen@redhat.com>
When performing a rolling update do not try to create any new osds with `ceph-volume lvm batch`. This is troublesome because when upgrading to nautilus the devices list might contain devices that are currently being used by ceph-disk and have GPT headers on them, which will cause ceph-volume to fail when trying to use such a device. Any devices originally created by ceph-disk will need to be removed from the devices list before any new osds can be created. Signed-off-by: Andrew Schoen <aschoen@redhat.com>
We do this so that the ceph-config role can most accurately report the number of osds for the generation of the ceph.conf file. We don't want to use ceph-volume to determine the number of osds because in an upgrade to nautilus ceph-volume won't be able to accurately count osds created by ceph-disk. Signed-off-by: Andrew Schoen <aschoen@redhat.com>
Signed-off-by: Andrew Schoen <aschoen@redhat.com>
6ba3345
to
8203472
Compare
|
@guits I think I've addressed everything here, would you mind taking another look? I reran my test manually and it still succeeds as well. https://2.jenkins.ceph.com/job/ceph-ansible-scenario/572/consoleFull |
|
failures are unrelated, merging anyway |
When upgrading to nautlius run
ceph-volume simple scanandceph-volume simple activate --allto migrate any runningceph-disk osds to ceph-volume.
Closes: https://bugzilla.redhat.com/show_bug.cgi?id=1656460
Signed-off-by: Andrew Schoen aschoen@redhat.com