New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ceph-disk: add --filestore argument, default to --bluestore #15437

Merged
merged 1 commit into from Jun 7, 2017

Conversation

Projects
None yet
1 participant
@ghost

ghost commented Jun 2, 2017

  • rados
  • ceph-disk

@ghost ghost added bluestore feature labels Jun 2, 2017

@ghost ghost requested a review from liewegas Jun 2, 2017

@liewegas liewegas added the needs-qa label Jun 2, 2017

@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 2, 2017

pushed to ceph-ci to debug the workunit

ghost commented Jun 2, 2017

pushed to ceph-ci to debug the workunit

@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 3, 2017

@liewegas for rados do we need something else beside removing the experimental status (that was merged at 036470a) ?

ghost commented Jun 3, 2017

@liewegas for rados do we need something else beside removing the experimental status (that was merged at 036470a) ?

@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 3, 2017

https://shaman.ceph.com/builds/ceph/wip-bluestore/d7fe74b56cebdbde7c3811fdd70e6d17fad4964a/notcmalloc/45798/ failed with

[ 76%] Building CXX object src/test/msgr/CMakeFiles/ceph_test_async_driver.dir/test_async_driver.cc.o
c++: internal compiler error: Killed (program cc1plus)

repushing

ghost commented Jun 3, 2017

https://shaman.ceph.com/builds/ceph/wip-bluestore/d7fe74b56cebdbde7c3811fdd70e6d17fad4964a/notcmalloc/45798/ failed with

[ 76%] Building CXX object src/test/msgr/CMakeFiles/ceph_test_async_driver.dir/test_async_driver.cc.o
c++: internal compiler error: Killed (program cc1plus)

repushing

@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 3, 2017

100 - test-erasure-eio.sh (Failed)

jenkins test this please

ghost commented Jun 3, 2017

100 - test-erasure-eio.sh (Failed)

jenkins test this please

@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 3, 2017

https://shaman.ceph.com/builds/ceph/wip-bluestore/ad9bd66ea3873c91ad12e99b75f3c9408167ce31/default/45828/

[ 34%] Building CXX object src/mon/CMakeFiles/mon.dir/AuthMonitor.cc.o
{standard input}: Assembler messages:
{standard input}:586375: Warning: end of file not at end of a line; newline inserted
{standard input}: Error: open CFI at the end of file; missing .cfi_endproc directive
c++: internal compiler error: Killed (program cc1plus)
Please submit a full bug report,
with preprocessed source if appropriate.
See http://bugzilla.redhat.com/bugzilla for instructions.
make[2]: *** [src/mds/CMakeFiles/mds.dir/MDCache.cc.o] Error 4

looks like shaman on centos 7 is having issues, ignoring it

ghost commented Jun 3, 2017

https://shaman.ceph.com/builds/ceph/wip-bluestore/ad9bd66ea3873c91ad12e99b75f3c9408167ce31/default/45828/

[ 34%] Building CXX object src/mon/CMakeFiles/mon.dir/AuthMonitor.cc.o
{standard input}: Assembler messages:
{standard input}:586375: Warning: end of file not at end of a line; newline inserted
{standard input}: Error: open CFI at the end of file; missing .cfi_endproc directive
c++: internal compiler error: Killed (program cc1plus)
Please submit a full bug report,
with preprocessed source if appropriate.
See http://bugzilla.redhat.com/bugzilla for instructions.
make[2]: *** [src/mds/CMakeFiles/mds.dir/MDCache.cc.o] Error 4

looks like shaman on centos 7 is having issues, ignoring it

@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 3, 2017

teuthology-suite -k distro --verbose --suite ceph-disk --ceph wip-bluestore --distro ubuntu --filter-out centos --machine-type vps

ghost commented Jun 3, 2017

teuthology-suite -k distro --verbose --suite ceph-disk --ceph wip-bluestore --distro ubuntu --filter-out centos --machine-type vps
@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 3, 2017

100 - test-erasure-eio.sh (Failed)

jenkins test this please

ghost commented Jun 3, 2017

100 - test-erasure-eio.sh (Failed)

jenkins test this please

@liewegas

This comment has been minimized.

Show comment
Hide comment
@liewegas

liewegas Jun 5, 2017

Member

retest this please

Member

liewegas commented Jun 5, 2017

retest this please

@liewegas

This comment has been minimized.

Show comment
Hide comment
@liewegas

liewegas Jun 5, 2017

Member

retest this please

Member

liewegas commented Jun 5, 2017

retest this please

@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 5, 2017

I'll check the ceph-disk failures

ghost commented Jun 5, 2017

I'll check the ceph-disk failures

@liewegas

This comment has been minimized.

Show comment
Hide comment
@liewegas

liewegas Jun 5, 2017

Member
Member

liewegas commented Jun 5, 2017

@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 5, 2017

retest this please (this is getting old ;-)

ghost commented Jun 5, 2017

retest this please (this is getting old ;-)

@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 5, 2017

fixed test error on ceph-disk suite & repushed to ceph-ci

ghost commented Jun 5, 2017

fixed test error on ceph-disk suite & repushed to ceph-ci

@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 5, 2017

all ceph-disk suite pass except for the tests verifying backward compatibility pre-lockbox, because it does not know about --filestore.

ghost commented Jun 5, 2017

all ceph-disk suite pass except for the tests verifying backward compatibility pre-lockbox, because it does not know about --filestore.

@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 5, 2017

teuthology-suite -k distro --verbose --suite ceph-disk --ceph wip-bluestore --machine-type vps

ghost commented Jun 5, 2017

teuthology-suite -k distro --verbose --suite ceph-disk --ceph wip-bluestore --machine-type vps
@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 5, 2017

added missing --filestore in ceph-helpers.sh and repushed

ghost commented Jun 5, 2017

added missing --filestore in ceph-helpers.sh and repushed

@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 6, 2017

remove remaining --experimental=bluestore in tests, fix osd-dup.sh which created a filestore instead of a bluestore. repushed.

ghost commented Jun 6, 2017

remove remaining --experimental=bluestore in tests, fix osd-dup.sh which created a filestore instead of a bluestore. repushed.

@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost commented Jun 6, 2017

running a long live test to debug the ceph-disk suite failure http://pulpito.ceph.com/loic-2017-06-06_09:04:30-ceph-disk-wip-bluestore-distro-basic-vps/

@liewegas

This comment has been minimized.

Show comment
Hide comment
@liewegas

liewegas Jun 6, 2017

Member
Member

liewegas commented Jun 6, 2017

@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 6, 2017

fixed the ceph-disk suite bug (no lockbox compatibility did not create the "type" file and activation was failing). We should be good now ! Repushing to ceph-ci.

ghost commented Jun 6, 2017

fixed the ceph-disk suite bug (no lockbox compatibility did not create the "type" file and activation was failing). We should be good now ! Repushing to ceph-ci.

@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 6, 2017

note to self: once it passes, rebase to verify #14074 did not break things

ghost commented Jun 6, 2017

note to self: once it passes, rebase to verify #14074 did not break things

@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 6, 2017

========================================================= 23 passed, 1 skipped in 942.95 seconds =========================================================

ceph-disk suite passes (manually on Ubuntu 16.04) but I don't see why it would not pass on other OS

ghost commented Jun 6, 2017

========================================================= 23 passed, 1 skipped in 942.95 seconds =========================================================

ceph-disk suite passes (manually on Ubuntu 16.04) but I don't see why it would not pass on other OS

@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 6, 2017

teuthology-suite --priority 100 -k distro --verbose --suite ceph-disk --ceph wip-bluestore --machine-type vps

ghost commented Jun 6, 2017

teuthology-suite --priority 100 -k distro --verbose --suite ceph-disk --ceph wip-bluestore --machine-type vps
ceph-disk: add --filestore argument, default to --bluestore
Signed-off-by: Sage Weil <sage@redhat.com>
Signed-off-by: Loic Dachary <loic@dachary.org>
@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 6, 2017

rebased & repushed

ghost commented Jun 6, 2017

rebased & repushed

@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 6, 2017

vps host disk died, rescheduled a run of the ceph-disk suite

ghost commented Jun 6, 2017

vps host disk died, rescheduled a run of the ceph-disk suite

@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 6, 2017

I consistently get

2017-06-06T21:31:14.150 INFO:tasks.ceph.mgr.x.ovh017.stderr:/home/jenkins-build/build/workspace/ceph-dev-new-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos7/DIST/centos7/MACHINE_SIZE/huge/release/12.0.2-2357-gdd1f471/rpm/el7/BUILD/ceph-12.0.2-2357-gdd1f471/src/mon/PGMap.cc: In function 'void PGMap::apply_incremental(CephContext*, const PGMap::Incremental&)' thread 7f1ee3b94700 time 2017-06-06 21:31:14.130335
2017-06-06T21:31:14.150 INFO:tasks.ceph.mgr.x.ovh017.stderr:/home/jenkins-build/build/workspace/ceph-dev-new-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos7/DIST/centos7/MACHINE_SIZE/huge/release/12.0.2-2357-gdd1f471/rpm/el7/BUILD/ceph-12.0.2-2357-gdd1f471/src/mon/PGMap.cc: 1085: FAILED assert(osd_stat.size() == osd_epochs.size())
2017-06-06T21:31:14.151 INFO:tasks.ceph.mgr.x.ovh017.stderr: ceph version  12.0.2-2357-gdd1f471 (dd1f47126deb239f70a10183867ef6a20d611d69) luminous (dev)
2017-06-06T21:31:14.151 INFO:tasks.ceph.mgr.x.ovh017.stderr: 1: (ceph::__ceph_assert_fail(char const*, char const*, int, char const*)+0x110) [0x7f1eefc7b500]
2017-06-06T21:31:14.151 INFO:tasks.ceph.mgr.x.ovh017.stderr: 2: (PGMap::apply_incremental(CephContext*, PGMap::Incremental const&)+0xf10) [0x7f1eefaf24e0]
2017-06-06T21:31:14.152 INFO:tasks.ceph.mgr.x.ovh017.stderr: 3: (ClusterState::update_delta_stats()+0xa9) [0x7f1eefb2e459]
2017-06-06T21:31:14.152 INFO:tasks.ceph.mgr.x.ovh017.stderr: 4: (DaemonServer::send_report()+0xcd) [0x7f1eefb1f3ed]
2017-06-06T21:31:14.152 INFO:tasks.ceph.mgr.x.ovh017.stderr: 5: (Mgr::tick()+0x70) [0x7f1eefb51990]
2017-06-06T21:31:14.152 INFO:tasks.ceph.mgr.x.ovh017.stderr: 6: (MgrStandby::tick()+0x89) [0x7f1eefb49489]
2017-06-06T21:31:14.152 INFO:tasks.ceph.mgr.x.ovh017.stderr: 7: (FunctionContext::finish(int)+0x2a) [0x7f1eefb2a97a]
2017-06-06T21:31:14.152 INFO:tasks.ceph.mgr.x.ovh017.stderr: 8: (Context::complete(int)+0x9) [0x7f1eefb26939]
2017-06-06T21:31:14.152 INFO:tasks.ceph.mgr.x.ovh017.stderr: 9: (SafeTimer::timer_thread()+0x104) [0x7f1eefc77dc4]
2017-06-06T21:31:14.152 INFO:tasks.ceph.mgr.x.ovh017.stderr: 10: (SafeTimerThread::entry()+0xd) [0x7f1eefc797ed]
2017-06-06T21:31:14.152 INFO:tasks.ceph.mgr.x.ovh017.stderr: 11: (()+0x7dc5) [0x7f1eed5d8dc5]
2017-06-06T21:31:14.152 INFO:tasks.ceph.mgr.x.ovh017.stderr: 12: (clone()+0x6d) [0x7f1eec6bd73d]
2017-06-06T21:31:14.152 INFO:tasks.ceph.mgr.x.ovh017.stderr: NOTE: a copy of the executable, or `objdump -rdS ` is needed to interpret this.
2017-06-06T21:31:14.153 INFO:tasks.ceph.mgr.x.ovh017.stderr:2017-06-06 21:31:14.132289 7f1ee3b94700 -1 /home/jenkins-build/build/workspace/ceph-dev-new-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos7/DIST/centos7/MACHINE_SIZE/huge/release/12.0.2-2357-gdd1f471/rpm/el7/BUILD/ceph-12.0.2-2357-gdd1f471/src/mon/PGMap.cc: In function 'void PGMap::apply_incremental(CephContext*, const PGMap::Incremental&)' thread 7f1ee3b94700 time 2017-06-06 21:31:14.130335

but it turns out this has been happening all along and did not prevent the tests from running successfully. I've been chasing a ghost ;-)

ghost commented Jun 6, 2017

I consistently get

2017-06-06T21:31:14.150 INFO:tasks.ceph.mgr.x.ovh017.stderr:/home/jenkins-build/build/workspace/ceph-dev-new-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos7/DIST/centos7/MACHINE_SIZE/huge/release/12.0.2-2357-gdd1f471/rpm/el7/BUILD/ceph-12.0.2-2357-gdd1f471/src/mon/PGMap.cc: In function 'void PGMap::apply_incremental(CephContext*, const PGMap::Incremental&)' thread 7f1ee3b94700 time 2017-06-06 21:31:14.130335
2017-06-06T21:31:14.150 INFO:tasks.ceph.mgr.x.ovh017.stderr:/home/jenkins-build/build/workspace/ceph-dev-new-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos7/DIST/centos7/MACHINE_SIZE/huge/release/12.0.2-2357-gdd1f471/rpm/el7/BUILD/ceph-12.0.2-2357-gdd1f471/src/mon/PGMap.cc: 1085: FAILED assert(osd_stat.size() == osd_epochs.size())
2017-06-06T21:31:14.151 INFO:tasks.ceph.mgr.x.ovh017.stderr: ceph version  12.0.2-2357-gdd1f471 (dd1f47126deb239f70a10183867ef6a20d611d69) luminous (dev)
2017-06-06T21:31:14.151 INFO:tasks.ceph.mgr.x.ovh017.stderr: 1: (ceph::__ceph_assert_fail(char const*, char const*, int, char const*)+0x110) [0x7f1eefc7b500]
2017-06-06T21:31:14.151 INFO:tasks.ceph.mgr.x.ovh017.stderr: 2: (PGMap::apply_incremental(CephContext*, PGMap::Incremental const&)+0xf10) [0x7f1eefaf24e0]
2017-06-06T21:31:14.152 INFO:tasks.ceph.mgr.x.ovh017.stderr: 3: (ClusterState::update_delta_stats()+0xa9) [0x7f1eefb2e459]
2017-06-06T21:31:14.152 INFO:tasks.ceph.mgr.x.ovh017.stderr: 4: (DaemonServer::send_report()+0xcd) [0x7f1eefb1f3ed]
2017-06-06T21:31:14.152 INFO:tasks.ceph.mgr.x.ovh017.stderr: 5: (Mgr::tick()+0x70) [0x7f1eefb51990]
2017-06-06T21:31:14.152 INFO:tasks.ceph.mgr.x.ovh017.stderr: 6: (MgrStandby::tick()+0x89) [0x7f1eefb49489]
2017-06-06T21:31:14.152 INFO:tasks.ceph.mgr.x.ovh017.stderr: 7: (FunctionContext::finish(int)+0x2a) [0x7f1eefb2a97a]
2017-06-06T21:31:14.152 INFO:tasks.ceph.mgr.x.ovh017.stderr: 8: (Context::complete(int)+0x9) [0x7f1eefb26939]
2017-06-06T21:31:14.152 INFO:tasks.ceph.mgr.x.ovh017.stderr: 9: (SafeTimer::timer_thread()+0x104) [0x7f1eefc77dc4]
2017-06-06T21:31:14.152 INFO:tasks.ceph.mgr.x.ovh017.stderr: 10: (SafeTimerThread::entry()+0xd) [0x7f1eefc797ed]
2017-06-06T21:31:14.152 INFO:tasks.ceph.mgr.x.ovh017.stderr: 11: (()+0x7dc5) [0x7f1eed5d8dc5]
2017-06-06T21:31:14.152 INFO:tasks.ceph.mgr.x.ovh017.stderr: 12: (clone()+0x6d) [0x7f1eec6bd73d]
2017-06-06T21:31:14.152 INFO:tasks.ceph.mgr.x.ovh017.stderr: NOTE: a copy of the executable, or `objdump -rdS ` is needed to interpret this.
2017-06-06T21:31:14.153 INFO:tasks.ceph.mgr.x.ovh017.stderr:2017-06-06 21:31:14.132289 7f1ee3b94700 -1 /home/jenkins-build/build/workspace/ceph-dev-new-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos7/DIST/centos7/MACHINE_SIZE/huge/release/12.0.2-2357-gdd1f471/rpm/el7/BUILD/ceph-12.0.2-2357-gdd1f471/src/mon/PGMap.cc: In function 'void PGMap::apply_incremental(CephContext*, const PGMap::Incremental&)' thread 7f1ee3b94700 time 2017-06-06 21:31:14.130335

but it turns out this has been happening all along and did not prevent the tests from running successfully. I've been chasing a ghost ;-)

@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 6, 2017

I don't get why it's failing consistently now. Trying to run the suite manually once more to get a clue.

ghost commented Jun 6, 2017

I don't get why it's failing consistently now. Trying to run the suite manually once more to get a clue.

@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 6, 2017

Manual test works... can't figure out the difference with what happens when it is run to completion...

ghost commented Jun 6, 2017

Manual test works... can't figure out the difference with what happens when it is run to completion...

@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 7, 2017

Turns out the failure to run the ceph-disk suite comes from a bug on master http://tracker.ceph.com/issues/20208 . Rebasing on a version of master dated a few weeks ago that passes http://pulpito.ceph.com/loic-2017-06-07_07:30:34-ceph-disk-wip-19941-ceph-disk-distro-basic-vps

ghost commented Jun 7, 2017

Turns out the failure to run the ceph-disk suite comes from a bug on master http://tracker.ceph.com/issues/20208 . Rebasing on a version of master dated a few weeks ago that passes http://pulpito.ceph.com/loic-2017-06-07_07:30:34-ceph-disk-wip-19941-ceph-disk-distro-basic-vps

@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 7, 2017

teuthology-suite --priority 10 -k distro --verbose --suite ceph-disk --ceph wip-bluestore-old-base --machine-type vps

ghost commented Jun 7, 2017

teuthology-suite --priority 10 -k distro --verbose --suite ceph-disk --ceph wip-bluestore-old-base --machine-type vps
@ghost

This comment has been minimized.

Show comment
Hide comment
@ghost

ghost Jun 7, 2017

@liewegas I believe it is ready to merge. The ceph-disk suite passes now (although using an older version of master from a few weeks ago because of http://tracker.ceph.com/issues/20208).

ghost commented Jun 7, 2017

@liewegas I believe it is ready to merge. The ceph-disk suite passes now (although using an older version of master from a few weeks ago because of http://tracker.ceph.com/issues/20208).

@liewegas liewegas merged commit fca1721 into ceph:master Jun 7, 2017

3 checks passed

Signed-off-by all commits in this PR are signed
Details
Unmodifed Submodules submodules for project are unmodified
Details
default Build finished.
Details
@liewegas

This comment has been minimized.

Show comment
Hide comment
@liewegas

liewegas Jun 7, 2017

Member

yay!

Member

liewegas commented Jun 7, 2017

yay!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment