diff options
author | Qiaowei Ren <qiaowei.ren@intel.com> | 2018-03-01 14:38:11 +0800 |
---|---|---|
committer | Qiaowei Ren <qiaowei.ren@intel.com> | 2018-03-01 14:38:11 +0800 |
commit | 7da45d65be36d36b880cc55c5036e96c24b53f00 (patch) | |
tree | d4f944eb4f8f8de50a9a7584ffa408dc3a3185b2 /src/ceph/qa/suites/upgrade | |
parent | 691462d09d0987b47e112d6ee8740375df3c51b2 (diff) |
remove ceph code
This patch removes initial ceph code, due to license issue.
Change-Id: I092d44f601cdf34aed92300fe13214925563081c
Signed-off-by: Qiaowei Ren <qiaowei.ren@intel.com>
Diffstat (limited to 'src/ceph/qa/suites/upgrade')
268 files changed, 0 insertions, 3402 deletions
diff --git a/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/basic/% b/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/basic/% deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/basic/% +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/basic/0-cluster/start.yaml b/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/basic/0-cluster/start.yaml deleted file mode 100644 index ea9c37d..0000000 --- a/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/basic/0-cluster/start.yaml +++ /dev/null @@ -1,14 +0,0 @@ -roles: -- - mon.a - - osd.0 - - osd.1 -- - mon.b - - mon.c - - osd.2 - - osd.3 -- - client.0 -overrides: - ceph: - log-whitelist: - - failed to encode map - fs: xfs diff --git a/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/basic/1-install/hammer-client-x.yaml b/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/basic/1-install/hammer-client-x.yaml deleted file mode 100644 index ffd4194..0000000 --- a/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/basic/1-install/hammer-client-x.yaml +++ /dev/null @@ -1,11 +0,0 @@ -tasks: -- install: - branch: hammer - exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] -- print: "**** done install hammer" -upgrade_workload: - sequential: - - install.upgrade: - exclude_packages: ['ceph-test-dbg'] - client.0: - - print: "**** done install.upgrade client.0" diff --git a/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/basic/2-workload/rbd_api_tests.yaml b/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/basic/2-workload/rbd_api_tests.yaml deleted file mode 100644 index 6638d14..0000000 --- a/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/basic/2-workload/rbd_api_tests.yaml +++ /dev/null @@ -1,26 +0,0 @@ -overrides: - ceph: - conf: - client: - rbd default features: 13 -tasks: -- exec: - client.0: - - "cp $(which ceph_test_librbd_api) $TESTDIR/ceph_test_librbd_api" -- sequential: - - upgrade_workload -- ceph: -- print: "**** done ceph" -- exec: - client.0: - - "cp --force $TESTDIR/ceph_test_librbd_api $(which ceph_test_librbd_api)" - - "rm -rf $TESTDIR/ceph_test_librbd_api" -- print: "**** done reverting to hammer ceph_test_librbd_api" -- workunit: - branch: hammer - clients: - client.0: - - rbd/test_librbd_api.sh - env: - RBD_FEATURES: "13" -- print: "**** done rbd/test_librbd_api.sh" diff --git a/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/basic/2-workload/rbd_cli_import_export.yaml b/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/basic/2-workload/rbd_cli_import_export.yaml deleted file mode 100644 index dfaa0e8..0000000 --- a/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/basic/2-workload/rbd_cli_import_export.yaml +++ /dev/null @@ -1,13 +0,0 @@ -tasks: -- sequential: - - upgrade_workload -- ceph: -- print: "**** done ceph" -- workunit: - branch: hammer - clients: - client.0: - - rbd/import_export.sh - env: - RBD_CREATE_ARGS: --image-feature layering,exclusive-lock,object-map -- print: "**** done rbd/import_export.sh" diff --git a/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/rbd/% b/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/rbd/% deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/rbd/% +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/rbd/0-cluster/start.yaml b/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/rbd/0-cluster/start.yaml deleted file mode 100644 index 4c9f324..0000000 --- a/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/rbd/0-cluster/start.yaml +++ /dev/null @@ -1,17 +0,0 @@ -roles: -- - mon.a - - mon.b - - mon.c - - osd.0 - - osd.1 - - osd.2 - - client.0 -- - client.1 -overrides: - ceph: - log-whitelist: - - failed to encode map - fs: xfs - conf: - client: - rbd default features: 1 diff --git a/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/rbd/1-install/hammer-client-x.yaml b/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/rbd/1-install/hammer-client-x.yaml deleted file mode 100644 index a625642..0000000 --- a/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/rbd/1-install/hammer-client-x.yaml +++ /dev/null @@ -1,11 +0,0 @@ -tasks: -- install: - branch: hammer - exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] -- print: "**** done install hammer" -- install.upgrade: - exclude_packages: ['ceph-test-dbg'] - client.1: -- print: "**** done install.upgrade client.1" -- ceph: -- print: "**** done ceph" diff --git a/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/rbd/2-workload/rbd_notification_tests.yaml b/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/rbd/2-workload/rbd_notification_tests.yaml deleted file mode 100644 index 984dfa0..0000000 --- a/src/ceph/qa/suites/upgrade/client-upgrade/hammer-client-x/rbd/2-workload/rbd_notification_tests.yaml +++ /dev/null @@ -1,21 +0,0 @@ -tasks: -- workunit: - branch: hammer - clients: - client.0: - - rbd/notify_master.sh - client.1: - - rbd/notify_slave.sh - env: - RBD_FEATURES: "13" -- print: "**** done rbd: old librbd -> new librbd" -- workunit: - branch: hammer - clients: - client.0: - - rbd/notify_slave.sh - client.1: - - rbd/notify_master.sh - env: - RBD_FEATURES: "13" -- print: "**** done rbd: new librbd -> old librbd" diff --git a/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/basic/% b/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/basic/% deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/basic/% +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/basic/0-cluster/start.yaml b/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/basic/0-cluster/start.yaml deleted file mode 100644 index a4cd754..0000000 --- a/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/basic/0-cluster/start.yaml +++ /dev/null @@ -1,13 +0,0 @@ -roles: -- - mon.a - - mon.b - - mon.c - - osd.0 - - osd.1 - - osd.2 -- - client.0 -overrides: - ceph: - log-whitelist: - - failed to encode map - fs: xfs diff --git a/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/basic/1-install/jewel-client-x.yaml b/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/basic/1-install/jewel-client-x.yaml deleted file mode 100644 index 87ea402..0000000 --- a/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/basic/1-install/jewel-client-x.yaml +++ /dev/null @@ -1,11 +0,0 @@ -tasks: -- install: - branch: jewel - exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] -- print: "**** done install jewel" -upgrade_workload: - sequential: - - install.upgrade: - exclude_packages: ['ceph-test', 'ceph-test-dbg'] - client.0: - - print: "**** done install.upgrade to -x on client.0" diff --git a/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/basic/2-workload/rbd_api_tests.yaml b/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/basic/2-workload/rbd_api_tests.yaml deleted file mode 100644 index 8939f3a..0000000 --- a/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/basic/2-workload/rbd_api_tests.yaml +++ /dev/null @@ -1,21 +0,0 @@ -tasks: -- exec: - client.0: - - "cp $(which ceph_test_librbd_api) $TESTDIR/ceph_test_librbd_api" -- sequential: - - upgrade_workload -- ceph: -- print: "**** done ceph" -- exec: - client.0: - - "cp --force $TESTDIR/ceph_test_librbd_api $(which ceph_test_librbd_api)" - - "rm -rf $TESTDIR/ceph_test_librbd_api" -- print: "**** done reverting to jewel ceph_test_librbd_api" -- workunit: - branch: kraken - clients: - client.0: - - rbd/test_librbd_api.sh - env: - RBD_FEATURES: "13" -- print: "**** done rbd/test_librbd_api.sh" diff --git a/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/basic/2-workload/rbd_cli_import_export.yaml b/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/basic/2-workload/rbd_cli_import_export.yaml deleted file mode 100644 index 545354f..0000000 --- a/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/basic/2-workload/rbd_cli_import_export.yaml +++ /dev/null @@ -1,13 +0,0 @@ -tasks: -- sequential: - - upgrade_workload -- ceph: -- print: "**** done ceph" -- workunit: - branch: jewel - clients: - client.0: - - rbd/import_export.sh - env: - RBD_CREATE_ARGS: --image-feature layering,exclusive-lock,object-map -- print: "**** done rbd/import_export.sh" diff --git a/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/rbd/% b/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/rbd/% deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/rbd/% +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/rbd/0-cluster/start.yaml b/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/rbd/0-cluster/start.yaml deleted file mode 100644 index 4db664b..0000000 --- a/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/rbd/0-cluster/start.yaml +++ /dev/null @@ -1,14 +0,0 @@ -roles: -- - mon.a - - mon.b - - mon.c - - osd.0 - - osd.1 - - osd.2 - - client.0 -- - client.1 -overrides: - ceph: - log-whitelist: - - failed to encode map - fs: xfs diff --git a/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/rbd/1-install/jewel-client-x.yaml b/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/rbd/1-install/jewel-client-x.yaml deleted file mode 100644 index 4ce73a4..0000000 --- a/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/rbd/1-install/jewel-client-x.yaml +++ /dev/null @@ -1,11 +0,0 @@ -tasks: -- install: - branch: jewel - exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] -- print: "**** done install jewel" -- install.upgrade: - exclude_packages: ['ceph-test', 'ceph-test-dbg'] - client.1: -- print: "**** done install.upgrade to -x on client.0" -- ceph: -- print: "**** done ceph task" diff --git a/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/rbd/2-features/defaults.yaml b/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/rbd/2-features/defaults.yaml deleted file mode 100644 index dff6623..0000000 --- a/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/rbd/2-features/defaults.yaml +++ /dev/null @@ -1,6 +0,0 @@ -overrides: - ceph: - conf: - client: - rbd default features: 61 - diff --git a/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/rbd/2-features/layering.yaml b/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/rbd/2-features/layering.yaml deleted file mode 100644 index 5613d01..0000000 --- a/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/rbd/2-features/layering.yaml +++ /dev/null @@ -1,6 +0,0 @@ -overrides: - ceph: - conf: - client: - rbd default features: 1 - diff --git a/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/rbd/3-workload/rbd_notification_tests.yaml b/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/rbd/3-workload/rbd_notification_tests.yaml deleted file mode 100644 index 1fb6822..0000000 --- a/src/ceph/qa/suites/upgrade/client-upgrade/jewel-client-x/rbd/3-workload/rbd_notification_tests.yaml +++ /dev/null @@ -1,21 +0,0 @@ -tasks: -- workunit: - branch: jewel - clients: - client.0: - - rbd/notify_master.sh - client.1: - - rbd/notify_slave.sh - env: - RBD_FEATURES: "13" -- print: "**** done rbd: old librbd -> new librbd" -- workunit: - branch: jewel - clients: - client.0: - - rbd/notify_slave.sh - client.1: - - rbd/notify_master.sh - env: - RBD_FEATURES: "13" -- print: "**** done rbd: new librbd -> old librbd" diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/% b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/% deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/% +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/0-cluster/start.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/0-cluster/start.yaml deleted file mode 100644 index bbddfb3..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/0-cluster/start.yaml +++ /dev/null @@ -1,21 +0,0 @@ -overrides: - ceph: - conf: - mon: - mon warn on legacy crush tunables: false - mon debug unsafe allow tier with nonempty snaps: true - log-whitelist: - - but it is still running - - wrongly marked me down - - reached quota -roles: -- - mon.a - - osd.0 - - osd.1 - - mgr.x -- - mon.b - - mon.c - - osd.2 - - osd.3 -- - client.0 - - client.1 diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/1-hammer-jewel-install/hammer-jewel.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/1-hammer-jewel-install/hammer-jewel.yaml deleted file mode 100644 index c57e071..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/1-hammer-jewel-install/hammer-jewel.yaml +++ /dev/null @@ -1,20 +0,0 @@ -tasks: -- install: - branch: hammer - exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] -- print: "**** done hammer" -- ceph: - fs: xfs - skip_mgr_daemons: true - add_osds_to_crush: true -- install.upgrade: - exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] - osd.0: - branch: jewel - osd.2: - branch: jewel -- print: "*** client.0 upgraded packages to jewel" -- parallel: - - workload - - upgrade-sequence -- print: "**** done parallel" diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/2-workload/+ b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/2-workload/+ deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/2-workload/+ +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/2-workload/ec-rados-default.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/2-workload/ec-rados-default.yaml deleted file mode 100644 index e4f3ee1..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/2-workload/ec-rados-default.yaml +++ /dev/null @@ -1,20 +0,0 @@ -workload: - full_sequential: - - rados: - clients: [client.0] - ops: 4000 - objects: 50 - ec_pool: true - write_append_excl: false - op_weights: - read: 100 - write: 0 - append: 100 - delete: 50 - snap_create: 50 - snap_remove: 50 - rollback: 50 - copy_from: 50 - setattr: 25 - rmattr: 25 - - print: "**** done rados ec task" diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/2-workload/rados_api.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/2-workload/rados_api.yaml deleted file mode 100644 index d86c2d2..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/2-workload/rados_api.yaml +++ /dev/null @@ -1,8 +0,0 @@ -workload: - full_sequential: - - workunit: - branch: hammer - clients: - client.0: - - cls - - print: "**** done cls 2-workload" diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/2-workload/rados_loadgenbig.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/2-workload/rados_loadgenbig.yaml deleted file mode 100644 index 50ba808..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/2-workload/rados_loadgenbig.yaml +++ /dev/null @@ -1,8 +0,0 @@ -workload: - full_sequential: - - workunit: - branch: hammer - clients: - client.0: - - rados/load-gen-big.sh - - print: "**** done rados/load-gen-big.sh 2-workload" diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/2-workload/test_rbd_api.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/2-workload/test_rbd_api.yaml deleted file mode 100644 index 997f7ba..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/2-workload/test_rbd_api.yaml +++ /dev/null @@ -1,8 +0,0 @@ -workload: - full_sequential: - - workunit: - branch: hammer - clients: - client.0: - - rbd/test_librbd.sh - - print: "**** done rbd/test_librbd.sh 2-workload" diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/2-workload/test_rbd_python.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/2-workload/test_rbd_python.yaml deleted file mode 100644 index d1046da..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/2-workload/test_rbd_python.yaml +++ /dev/null @@ -1,8 +0,0 @@ -workload: - full_sequential: - - workunit: - branch: hammer - clients: - client.0: - - rbd/test_librbd_python.sh - - print: "**** done rbd/test_librbd_python.sh 2-workload" diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/3-upgrade-sequence/upgrade-all.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/3-upgrade-sequence/upgrade-all.yaml deleted file mode 100644 index 1aaeac8..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/3-upgrade-sequence/upgrade-all.yaml +++ /dev/null @@ -1,18 +0,0 @@ -upgrade-sequence: - sequential: - - ceph.restart: - daemons: [osd.0, osd.1, osd.2, osd.3] - wait-for-healthy: false - wait-for-osds-up: true - - ceph.restart: - daemons: [mon.a, mon.b, mon.c] - wait-for-healthy: false - wait-for-osds-up: true - - print: "**** done ceph.restart do not wait for healthy" - - exec: - mon.a: - - sleep 300 # http://tracker.ceph.com/issues/17808 - - ceph osd set sortbitwise - - ceph osd set require_jewel_osds - - ceph.healthy: - - print: "**** done ceph.healthy" diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/3-upgrade-sequence/upgrade-osd-mds-mon.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/3-upgrade-sequence/upgrade-osd-mds-mon.yaml deleted file mode 100644 index f2093da..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/3-upgrade-sequence/upgrade-osd-mds-mon.yaml +++ /dev/null @@ -1,36 +0,0 @@ -upgrade-sequence: - sequential: - - ceph.restart: - daemons: [osd.0, osd.1] - wait-for-healthy: true - - sleep: - duration: 60 - - ceph.restart: - daemons: [osd.2, osd.3] - wait-for-healthy: true - - sleep: - duration: 60 - - ceph.restart: - daemons: [mon.a] - wait-for-healthy: false - - sleep: - duration: 60 - - print: "**** running mixed versions of osds and mons" -#do we need to use "ceph osd crush tunables hammer" ? - - exec: - mon.b: - - sudo ceph osd crush tunables hammer - - print: "**** done ceph osd crush tunables hammer" - - ceph.restart: - daemons: [mon.b, mon.c] - wait-for-healthy: false - - sleep: - duration: 30 - - exec: - osd.0: - - sleep 300 # http://tracker.ceph.com/issues/17808 - - ceph osd set sortbitwise - - ceph osd set require_jewel_osds - - ceph.healthy: - - sleep: - duration: 60 diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/3.5-finish.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/3.5-finish.yaml deleted file mode 100644 index 60a3cb6..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/3.5-finish.yaml +++ /dev/null @@ -1,5 +0,0 @@ -tasks: -- install.upgrade: - exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] - client.0: - branch: jewel diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/4-jewel.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/4-jewel.yaml deleted file mode 120000 index 987c18c..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/4-jewel.yaml +++ /dev/null @@ -1 +0,0 @@ -../../../../releases/jewel.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/5-hammer-jewel-x-upgrade/hammer-jewel-x.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/5-hammer-jewel-x-upgrade/hammer-jewel-x.yaml deleted file mode 100644 index ab41db6..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/5-hammer-jewel-x-upgrade/hammer-jewel-x.yaml +++ /dev/null @@ -1,14 +0,0 @@ -tasks: - - install.upgrade: - exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] - client.0: - branch: jewel - - print: "**** done install.upgrade client.0 to jewel" - - install.upgrade: - osd.0: - osd.2: - - print: "**** done install.upgrade daemons to x" - - parallel: - - workload2 - - upgrade-sequence2 - - print: "**** done parallel workload2 and upgrade-sequence2" diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/6-workload/+ b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/6-workload/+ deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/6-workload/+ +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/6-workload/ec-rados-default.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/6-workload/ec-rados-default.yaml deleted file mode 100644 index 9818541..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/6-workload/ec-rados-default.yaml +++ /dev/null @@ -1,29 +0,0 @@ -meta: -- desc: | - run run randomized correctness test for rados operations - on an erasure-coded pool -workload2: - full_sequential: - - rados: - erasure_code_profile: - name: teuthologyprofile2 - k: 2 - m: 1 - crush-failure-domain: osd - clients: [client.0] - ops: 4000 - objects: 50 - ec_pool: true - write_append_excl: false - op_weights: - read: 100 - write: 0 - append: 100 - delete: 50 - snap_create: 50 - snap_remove: 50 - rollback: 50 - copy_from: 50 - setattr: 25 - rmattr: 25 - - print: "**** done rados ec task" diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/6-workload/rados_api.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/6-workload/rados_api.yaml deleted file mode 100644 index 088976b..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/6-workload/rados_api.yaml +++ /dev/null @@ -1,11 +0,0 @@ -meta: -- desc: | - object class functional tests -workload2: - full_sequential: - - workunit: - branch: jewel - clients: - client.0: - - cls - - print: "**** done cls 2-workload" diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/6-workload/rados_loadgenbig.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/6-workload/rados_loadgenbig.yaml deleted file mode 100644 index 30f1307..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/6-workload/rados_loadgenbig.yaml +++ /dev/null @@ -1,11 +0,0 @@ -meta: -- desc: | - generate read/write load with rados objects ranging from 1MB to 25MB -workload2: - full_sequential: - - workunit: - branch: jewel - clients: - client.0: - - rados/load-gen-big.sh - - print: "**** done rados/load-gen-big.sh 2-workload" diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/6-workload/test_rbd_api.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/6-workload/test_rbd_api.yaml deleted file mode 100644 index e21839b..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/6-workload/test_rbd_api.yaml +++ /dev/null @@ -1,11 +0,0 @@ -meta: -- desc: | - librbd C and C++ api tests -workload2: - full_sequential: - - workunit: - branch: jewel - clients: - client.0: - - rbd/test_librbd.sh - - print: "**** done rbd/test_librbd.sh 2-workload" diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/6-workload/test_rbd_python.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/6-workload/test_rbd_python.yaml deleted file mode 100644 index cae2c06..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/6-workload/test_rbd_python.yaml +++ /dev/null @@ -1,11 +0,0 @@ -meta: -- desc: | - librbd python api tests -workload2: - full_sequential: - - workunit: - branch: jewel - clients: - client.0: - - rbd/test_librbd_python.sh - - print: "**** done rbd/test_librbd_python.sh 2-workload" diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/7-upgrade-sequence/upgrade-all.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/7-upgrade-sequence/upgrade-all.yaml deleted file mode 100644 index 356f8ad..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/7-upgrade-sequence/upgrade-all.yaml +++ /dev/null @@ -1,10 +0,0 @@ -meta: -- desc: | - upgrade the ceph cluster -upgrade-sequence2: - sequential: - - ceph.restart: - daemons: [mon.a, mon.b, mon.c, osd.0, osd.1, osd.2, osd.3] - wait-for-healthy: false - wait-for-osds-up: true - - print: "**** done ceph.restart all" diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/7-upgrade-sequence/upgrade-by-daemon.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/7-upgrade-sequence/upgrade-by-daemon.yaml deleted file mode 100644 index 0a69a7f..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/7-upgrade-sequence/upgrade-by-daemon.yaml +++ /dev/null @@ -1,30 +0,0 @@ -meta: -- desc: | - upgrade the ceph cluster, - upgrate in two steps - step one ordering: mon.a, mon.b, mon.c, osd.0, osd.1 - step two ordering: osd.2, osd.3 - ceph expected to be healthy state after each step -upgrade-sequence2: - sequential: - - ceph.restart: - daemons: [mon.a, mon.b, mon.c] - wait-for-healthy: true - - sleep: - duration: 60 - - ceph.restart: - daemons: [osd.0, osd.1] - wait-for-healthy: true - - sleep: - duration: 60 - - print: "**** running mixed versions of osds and mons" - - exec: - mon.b: - - sudo ceph osd crush tunables jewel - - print: "**** done ceph osd crush tunables jewel" - - ceph.restart: - daemons: [osd.2, osd.3] - wait-for-healthy: false - wait-for-osds-up: true - - sleep: - duration: 60 diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/8-luminous.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/8-luminous.yaml deleted file mode 120000 index 5283ac7..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/8-luminous.yaml +++ /dev/null @@ -1 +0,0 @@ -../../../../releases/luminous.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/9-final-workload/+ b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/9-final-workload/+ deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/9-final-workload/+ +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/9-final-workload/rados-snaps-few-objects.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/9-final-workload/rados-snaps-few-objects.yaml deleted file mode 100644 index e0b0ba1..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/9-final-workload/rados-snaps-few-objects.yaml +++ /dev/null @@ -1,13 +0,0 @@ -tasks: -- rados: - clients: [client.1] - ops: 4000 - objects: 50 - op_weights: - read: 100 - write: 100 - delete: 50 - snap_create: 50 - snap_remove: 50 - rollback: 50 -- print: "**** done 7-final-workload/rados-snaps-few-objects.yaml" diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/9-final-workload/rados_loadgenmix.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/9-final-workload/rados_loadgenmix.yaml deleted file mode 100644 index b1c6791..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/9-final-workload/rados_loadgenmix.yaml +++ /dev/null @@ -1,6 +0,0 @@ -tasks: - - workunit: - clients: - client.1: - - rados/load-gen-mix.sh - - print: "**** done 7-final-workload/rados_loadgenmix.yaml" diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/9-final-workload/rados_mon_thrash.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/9-final-workload/rados_mon_thrash.yaml deleted file mode 100644 index 807afb9..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/9-final-workload/rados_mon_thrash.yaml +++ /dev/null @@ -1,11 +0,0 @@ -tasks: - - sequential: - - mon_thrash: - revive_delay: 20 - thrash_delay: 1 - - workunit: - branch: jewel - clients: - client.1: - - rados/test-upgrade-v11.0.0.sh - - print: "**** done rados/test-upgrade-v11.0.0.sh from 7-final-workload" diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/9-final-workload/rbd_cls.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/9-final-workload/rbd_cls.yaml deleted file mode 100644 index 973c438..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/9-final-workload/rbd_cls.yaml +++ /dev/null @@ -1,6 +0,0 @@ -tasks: -- workunit: - clients: - client.1: - - cls/test_cls_rbd.sh -- print: "**** done 7-final-workload/rbd_cls.yaml" diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/9-final-workload/rbd_import_export.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/9-final-workload/rbd_import_export.yaml deleted file mode 100644 index d8116a9..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/9-final-workload/rbd_import_export.yaml +++ /dev/null @@ -1,8 +0,0 @@ -tasks: -- workunit: - clients: - client.1: - - rbd/import_export.sh - env: - RBD_CREATE_ARGS: --new-format -- print: "**** done rbd/import_export.sh from 7-final-workload" diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/9-final-workload/rgw_s3tests.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/9-final-workload/rgw_s3tests.yaml deleted file mode 100644 index f1cf2de..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/9-final-workload/rgw_s3tests.yaml +++ /dev/null @@ -1,11 +0,0 @@ -tasks: -- rgw: [client.1] -- s3tests: - client.1: - rgw_server: client.1 -- print: "**** done rgw_server from 7-final-workload" -overrides: - ceph: - conf: - client: - rgw lc debug interval: 10 diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/distros b/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/distros deleted file mode 120000 index ca99fee..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/parallel/distros +++ /dev/null @@ -1 +0,0 @@ -../../../../distros/supported/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/% b/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/% deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/% +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/0-cluster b/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/0-cluster deleted file mode 120000 index 9bb7a0d..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/0-cluster +++ /dev/null @@ -1 +0,0 @@ -../../jewel-x/stress-split/0-cluster
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/1-hammer-install-and-upgrade-to-jewel/hammer-to-jewel.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/1-hammer-install-and-upgrade-to-jewel/hammer-to-jewel.yaml deleted file mode 100644 index 212b8ff..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/1-hammer-install-and-upgrade-to-jewel/hammer-to-jewel.yaml +++ /dev/null @@ -1,83 +0,0 @@ -tasks: -- install: - branch: hammer - exclude_packages: - - ceph-mgr - - libcephfs2 - - libcephfs-devel - - libcephfs-dev -- print: '**** done hammer' -- ceph: - fs: xfs - skip_mgr_daemons: true - add_osds_to_crush: true -- install.upgrade: - exclude_packages: - - ceph-mgr - - libcephfs2 - - libcephfs-devel - - libcephfs-dev - osd.0: - branch: jewel - osd.3: - branch: jewel -- print: '*** client.0 upgraded packages to jewel' -- parallel: - - workload-h-j - - upgrade-sequence-h-j -- print: '**** done parallel' -- install.upgrade: - client.0: - branch: jewel - exclude_packages: - - ceph-mgr - - libcephfs2 - - libcephfs-devel - - libcephfs-dev -- exec: - osd.0: - - ceph osd set sortbitwise - - ceph osd set require_jewel_osds - - for p in `ceph osd pool ls` ; do ceph osd pool set $p use_gmt_hitset true ; - done -- print: '**** done install.upgrade client.0 to jewel' -upgrade-sequence-h-j: - sequential: - - ceph.restart: - daemons: - - osd.0 - - osd.1 - - osd.2 - - osd.3 - - osd.4 - - osd.5 - wait-for-healthy: false - wait-for-osds-up: true - - ceph.restart: - daemons: - - mon.a - - mon.b - - mon.c - wait-for-healthy: false - wait-for-osds-up: true - - print: '**** done ceph.restart do not wait for healthy' - - exec: - mon.a: - - sleep 300 - - ceph osd set require_jewel_osds - - ceph.healthy: null - - print: '**** done ceph.healthy' -workload-h-j: - full_sequential: - - workunit: - branch: hammer - clients: - client.0: - - cls - - print: "**** done cls 2-workload" - - workunit: - branch: hammer - clients: - client.0: - - rbd/test_librbd.sh - - print: "**** done rbd/test_librbd.sh 2-workload" diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/2-partial-upgrade b/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/2-partial-upgrade deleted file mode 120000 index fad7148..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/2-partial-upgrade +++ /dev/null @@ -1 +0,0 @@ -../../jewel-x/stress-split/2-partial-upgrade/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/3-thrash b/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/3-thrash deleted file mode 120000 index 894fdeb..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/3-thrash +++ /dev/null @@ -1 +0,0 @@ -../../jewel-x/stress-split/3-thrash/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/4-workload b/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/4-workload deleted file mode 120000 index 6135fb0..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/4-workload +++ /dev/null @@ -1 +0,0 @@ -../../jewel-x/stress-split/4-workload
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/5-finish-upgrade.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/5-finish-upgrade.yaml deleted file mode 120000 index 7d39ac6..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/5-finish-upgrade.yaml +++ /dev/null @@ -1 +0,0 @@ -../../jewel-x/stress-split/5-finish-upgrade.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/6-luminous.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/6-luminous.yaml deleted file mode 120000 index 5283ac7..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/6-luminous.yaml +++ /dev/null @@ -1 +0,0 @@ -../../../../releases/luminous.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/7-final-workload b/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/7-final-workload deleted file mode 120000 index 97adf26..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/7-final-workload +++ /dev/null @@ -1 +0,0 @@ -../../jewel-x/stress-split/7-final-workload/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/distros b/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/distros deleted file mode 120000 index ca99fee..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/stress-split/distros +++ /dev/null @@ -1 +0,0 @@ -../../../../distros/supported/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/tiering/% b/src/ceph/qa/suites/upgrade/hammer-jewel-x/tiering/% deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/tiering/% +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/tiering/0-cluster/start.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/tiering/0-cluster/start.yaml deleted file mode 100644 index 9cd743c..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/tiering/0-cluster/start.yaml +++ /dev/null @@ -1,17 +0,0 @@ -overrides: - ceph: - conf: - mon: - mon warn on legacy crush tunables: false - log-whitelist: - - but it is still running - - wrongly marked me down -roles: -- - mon.a - - osd.0 - - osd.1 -- - mon.b - - mon.c - - osd.2 - - osd.3 -- - client.0 diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/tiering/1-install-hammer-and-upgrade-to-jewel/hammer-to-jewel.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/tiering/1-install-hammer-and-upgrade-to-jewel/hammer-to-jewel.yaml deleted file mode 100644 index 7485dce..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/tiering/1-install-hammer-and-upgrade-to-jewel/hammer-to-jewel.yaml +++ /dev/null @@ -1,13 +0,0 @@ -tasks: -- install: - branch: hammer - exclude_packages: - - ceph-mgr - - libcephfs2 - - libcephfs-devel - - libcephfs-dev -- print: '**** done hammer' -- ceph: - fs: xfs - skip_mgr_daemons: true - add_osds_to_crush: true diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/tiering/2-setup-cache-tiering/% b/src/ceph/qa/suites/upgrade/hammer-jewel-x/tiering/2-setup-cache-tiering/% deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/tiering/2-setup-cache-tiering/% +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/tiering/2-setup-cache-tiering/0-create-base-tier/create-ec-pool.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/tiering/2-setup-cache-tiering/0-create-base-tier/create-ec-pool.yaml deleted file mode 100644 index f0e22bf..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/tiering/2-setup-cache-tiering/0-create-base-tier/create-ec-pool.yaml +++ /dev/null @@ -1,6 +0,0 @@ -tasks: -- exec: - client.0: - - ceph osd erasure-code-profile set t-profile crush-failure-domain=osd k=2 m=1 - - ceph osd pool create base-pool 4 4 erasure t-profile - - ceph osd pool application enable base-pool rados diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/tiering/2-setup-cache-tiering/0-create-base-tier/create-replicated-pool.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/tiering/2-setup-cache-tiering/0-create-base-tier/create-replicated-pool.yaml deleted file mode 100644 index 36dc06d..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/tiering/2-setup-cache-tiering/0-create-base-tier/create-replicated-pool.yaml +++ /dev/null @@ -1,5 +0,0 @@ -tasks: -- exec: - client.0: - - ceph osd pool create base-pool 4 - - ceph osd pool application enable base-pool rados diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/tiering/2-setup-cache-tiering/1-create-cache-tier.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/tiering/2-setup-cache-tiering/1-create-cache-tier.yaml deleted file mode 100644 index d9cc348..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/tiering/2-setup-cache-tiering/1-create-cache-tier.yaml +++ /dev/null @@ -1,14 +0,0 @@ -overrides: - ceph: - log-whitelist: - - must scrub before tier agent can activate -tasks: -- exec: - client.0: - - ceph osd pool create cache-pool 4 - - ceph osd tier add base-pool cache-pool - - ceph osd tier cache-mode cache-pool writeback - - ceph osd tier set-overlay base-pool cache-pool - - ceph osd pool set cache-pool hit_set_type bloom - - ceph osd pool set cache-pool hit_set_count 8 - - ceph osd pool set cache-pool hit_set_period 5 diff --git a/src/ceph/qa/suites/upgrade/hammer-jewel-x/tiering/3-upgrade.yaml b/src/ceph/qa/suites/upgrade/hammer-jewel-x/tiering/3-upgrade.yaml deleted file mode 100644 index b2fc171..0000000 --- a/src/ceph/qa/suites/upgrade/hammer-jewel-x/tiering/3-upgrade.yaml +++ /dev/null @@ -1,52 +0,0 @@ -tasks: -- parallel: - - workload - - upgrade-sequence -- print: "**** done parallel" - -workload: - sequential: - - rados: - clients: [client.0] - pools: [base-pool] - ops: 4000 - objects: 500 - op_weights: - read: 100 - write: 100 - delete: 50 - copy_from: 50 - cache_flush: 50 - cache_try_flush: 50 - cache_evict: 50 - - print: "**** done rados" - -upgrade-sequence: - sequential: - - install.upgrade: - exclude_packages: - - ceph-mgr - - libcephfs2 - - libcephfs-devel - - libcephfs-dev - osd.0: - branch: jewel - osd.2: - branch: jewel - - print: "*** done install.upgrade osd.0 and osd.2" - - ceph.restart: - daemons: [osd.0, osd.1, osd.2, osd.3] - wait-for-healthy: false - wait-for-osds-up: true - - ceph.restart: - daemons: [mon.a, mon.b, mon.c] - wait-for-healthy: false - wait-for-osds-up: true - - print: "**** done ceph.restart do not wait for healthy" - - exec: - mon.a: - - sleep 300 # http://tracker.ceph.com/issues/17808 - - ceph osd set sortbitwise - - ceph osd set require_jewel_osds - - ceph.healthy: - - print: "**** done ceph.healthy" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/ceph-deploy/% b/src/ceph/qa/suites/upgrade/jewel-x/ceph-deploy/% deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/ceph-deploy/% +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/jewel-x/ceph-deploy/distros/centos_latest.yaml b/src/ceph/qa/suites/upgrade/jewel-x/ceph-deploy/distros/centos_latest.yaml deleted file mode 120000 index b5973b9..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/ceph-deploy/distros/centos_latest.yaml +++ /dev/null @@ -1 +0,0 @@ -../../../../../distros/supported/centos_latest.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/jewel-x/ceph-deploy/distros/ubuntu_latest.yaml b/src/ceph/qa/suites/upgrade/jewel-x/ceph-deploy/distros/ubuntu_latest.yaml deleted file mode 120000 index cc5b15b..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/ceph-deploy/distros/ubuntu_latest.yaml +++ /dev/null @@ -1 +0,0 @@ -../../../../../distros/supported/ubuntu_latest.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/jewel-x/ceph-deploy/jewel-luminous.yaml b/src/ceph/qa/suites/upgrade/jewel-x/ceph-deploy/jewel-luminous.yaml deleted file mode 100644 index 9adede7..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/ceph-deploy/jewel-luminous.yaml +++ /dev/null @@ -1,82 +0,0 @@ -meta: -- desc: | - Setup 4 node ceph cluster using ceph-deploy, use latest - stable jewel as initial release, upgrade to luminous and - also setup mgr nodes along after upgrade, check for - cluster to reach healthy state, After upgrade run kernel tar/untar - task and systemd task. This test will detect any - ceph upgrade issue and systemd issues. -overrides: - ceph-deploy: - fs: xfs - conf: - global: - mon pg warn min per osd: 2 - osd: - osd pool default size: 2 - osd objectstore: filestore - osd sloppy crc: true - client: - rbd default features: 5 -openstack: -- machine: - disk: 100 -- volumes: - count: 3 - size: 30 -# reluctantely :( hard-coded machine type -# it will override command line args with teuthology-suite -machine_type: vps -roles: -- - mon.a - - mds.a - - osd.0 - - osd.1 - - osd.2 - - mgr.x -- - mon.b - - mgr.y -- - mon.c - - osd.3 - - osd.4 - - osd.5 -- - osd.6 - - osd.7 - - osd.8 - - client.0 -tasks: -- ssh-keys: -- print: "**** done ssh-keys" -- ceph-deploy: - branch: - stable: jewel - skip-mgr: True -- print: "**** done initial ceph-deploy" -- ceph-deploy.upgrade: - branch: - dev: luminous - setup-mgr-node: True - check-for-healthy: True - roles: - - mon.a - - mon.b - - mon.c - - osd.6 -- print: "**** done ceph-deploy upgrade" -- exec: - osd.0: - - ceph osd require-osd-release luminous - - ceph osd set-require-min-compat-client luminous -- print: "**** done `ceph osd require-osd-release luminous`" -- workunit: - clients: - all: - - kernel_untar_build.sh -- print: "**** done kernel_untar_build.sh" -- systemd: -- print: "**** done systemd" -- workunit: - clients: - all: - - rados/load-gen-mix.sh -- print: "**** done rados/load-gen-mix.sh" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/% b/src/ceph/qa/suites/upgrade/jewel-x/parallel/% deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/% +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/0-cluster/+ b/src/ceph/qa/suites/upgrade/jewel-x/parallel/0-cluster/+ deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/0-cluster/+ +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/0-cluster/openstack.yaml b/src/ceph/qa/suites/upgrade/jewel-x/parallel/0-cluster/openstack.yaml deleted file mode 100644 index f4d1349..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/0-cluster/openstack.yaml +++ /dev/null @@ -1,4 +0,0 @@ -openstack: - - volumes: # attached to each instance - count: 3 - size: 30 # GB diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/0-cluster/start.yaml b/src/ceph/qa/suites/upgrade/jewel-x/parallel/0-cluster/start.yaml deleted file mode 100644 index d1f1e10..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/0-cluster/start.yaml +++ /dev/null @@ -1,32 +0,0 @@ -meta: -- desc: | - Run ceph on two nodes, - with a separate client 0,1,2 third node. - Use xfs beneath the osds. - CephFS tests running on client 2,3 -roles: -- - mon.a - - mds.a - - mgr.x - - osd.0 - - osd.1 -- - mon.b - - mon.c - - osd.2 - - osd.3 -- - client.0 - - client.1 - - client.2 - - client.3 -- - client.4 -overrides: - ceph: - log-whitelist: - - scrub mismatch - - ScrubResult - - wrongly marked - - \(MDS_FAILED\) - - \(OBJECT_ - - is unresponsive - conf: - fs: xfs diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/1-jewel-install/jewel.yaml b/src/ceph/qa/suites/upgrade/jewel-x/parallel/1-jewel-install/jewel.yaml deleted file mode 100644 index c64b2cd..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/1-jewel-install/jewel.yaml +++ /dev/null @@ -1,60 +0,0 @@ -overrides: - ceph: - conf: - client.0: - debug ms: 1 - debug client: 10 - debug monc: 10 - client.1: - debug ms: 1 - debug client: 10 - debug monc: 10 - client.2: - debug ms: 1 - debug client: 10 - debug monc: 10 - client.3: - debug ms: 1 - debug client: 10 - debug monc: 10 -meta: -- desc: | - install ceph/jewel latest - run workload and upgrade-sequence in parallel - upgrade the client node -tasks: -- install: - branch: jewel - exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] -- print: "**** done installing jewel" -- ceph: - skip_mgr_daemons: true - add_osds_to_crush: true - log-whitelist: - - overall HEALTH_ - - \(FS_ - - \(MDS_ - - \(OSD_ - - \(MON_DOWN\) - - \(CACHE_POOL_ - - \(POOL_ - - \(MGR_DOWN\) - - \(PG_ - - Monitor daemon marked osd - - Behind on trimming - - is unresponsive - conf: - global: - mon warn on pool no app: false -- print: "**** done ceph" -- install.upgrade: - mon.a: - mon.b: -- print: "**** done install.upgrade mon.a and mon.b" -- parallel: - - workload - - upgrade-sequence -- print: "**** done parallel" -- install.upgrade: - client.0: -- print: "**** done install.upgrade on client.0" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/1.5-final-scrub.yaml b/src/ceph/qa/suites/upgrade/jewel-x/parallel/1.5-final-scrub.yaml deleted file mode 100644 index 83457c0..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/1.5-final-scrub.yaml +++ /dev/null @@ -1,11 +0,0 @@ -# do not require luminous osds at mkfs time; only set flag at -# the end of the test run, then do a final scrub (to convert any -# legacy snapsets), and verify we are healthy. -tasks: -- full_sequential_finally: - - ceph.osd_scrub_pgs: - cluster: ceph - - exec: - mon.a: - - ceph pg dump -f json-pretty - - "ceph pg dump sum -f json-pretty | grep num_legacy_snapsets | head -1 | grep ': 0'" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/2-workload/blogbench.yaml b/src/ceph/qa/suites/upgrade/jewel-x/parallel/2-workload/blogbench.yaml deleted file mode 100644 index 56eedbd..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/2-workload/blogbench.yaml +++ /dev/null @@ -1,14 +0,0 @@ -meta: -- desc: | - run a cephfs stress test - mount ceph-fuse on client.2 before running workunit -workload: - full_sequential: - - sequential: - - ceph-fuse: [client.2] - - print: "**** done ceph-fuse 2-workload" - - workunit: - clients: - client.2: - - suites/blogbench.sh - - print: "**** done suites/blogbench.sh 2-workload" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/2-workload/cache-pool-snaps.yaml b/src/ceph/qa/suites/upgrade/jewel-x/parallel/2-workload/cache-pool-snaps.yaml deleted file mode 100644 index dfbcbea..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/2-workload/cache-pool-snaps.yaml +++ /dev/null @@ -1,41 +0,0 @@ -overrides: - ceph: - log-whitelist: - - must scrub before tier agent can activate -tasks: -workload: - full_sequential: - - sequential: - - exec: - client.0: - - sudo ceph osd pool create base 4 - - sudo ceph osd pool create cache 4 - - sudo ceph osd tier add base cache - - sudo ceph osd tier cache-mode cache writeback - - sudo ceph osd tier set-overlay base cache - - sudo ceph osd pool set cache hit_set_type bloom - - sudo ceph osd pool set cache hit_set_count 8 - - sudo ceph osd pool set cache hit_set_period 3600 - - sudo ceph osd pool set cache target_max_objects 250 - - sudo ceph osd pool set cache min_read_recency_for_promote 0 - - sudo ceph osd pool set cache min_write_recency_for_promote 0 - - rados: - clients: [client.0] - pools: [base] - ops: 4000 - objects: 500 - pool_snaps: true - op_weights: - read: 100 - write: 100 - delete: 50 - copy_from: 50 - cache_flush: 50 - cache_try_flush: 50 - cache_evict: 50 - snap_create: 50 - snap_remove: 50 - rollback: 50 -openstack: - - machine: - ram: 15000 # MB diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/2-workload/ec-rados-default.yaml b/src/ceph/qa/suites/upgrade/jewel-x/parallel/2-workload/ec-rados-default.yaml deleted file mode 100644 index fb9d30f..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/2-workload/ec-rados-default.yaml +++ /dev/null @@ -1,24 +0,0 @@ -meta: -- desc: | - run run randomized correctness test for rados operations - on an erasure-coded pool -workload: - full_sequential: - - rados: - clients: [client.0] - ops: 4000 - objects: 50 - ec_pool: true - write_append_excl: false - op_weights: - read: 100 - write: 0 - append: 100 - delete: 50 - snap_create: 50 - snap_remove: 50 - rollback: 50 - copy_from: 50 - setattr: 25 - rmattr: 25 - - print: "**** done rados ec task" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/2-workload/rados_api.yaml b/src/ceph/qa/suites/upgrade/jewel-x/parallel/2-workload/rados_api.yaml deleted file mode 100644 index 348f1ae..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/2-workload/rados_api.yaml +++ /dev/null @@ -1,11 +0,0 @@ -meta: -- desc: | - object class functional tests -workload: - full_sequential: - - workunit: - branch: jewel - clients: - client.0: - - cls - - print: "**** done cls 2-workload" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/2-workload/test_rbd_api.yaml b/src/ceph/qa/suites/upgrade/jewel-x/parallel/2-workload/test_rbd_api.yaml deleted file mode 100644 index 15d892e..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/2-workload/test_rbd_api.yaml +++ /dev/null @@ -1,11 +0,0 @@ -meta: -- desc: | - librbd C and C++ api tests -workload: - full_sequential: - - workunit: - branch: jewel - clients: - client.0: - - rbd/test_librbd.sh - - print: "**** done rbd/test_librbd.sh 2-workload" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/2-workload/test_rbd_python.yaml b/src/ceph/qa/suites/upgrade/jewel-x/parallel/2-workload/test_rbd_python.yaml deleted file mode 100644 index bb2d3ea..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/2-workload/test_rbd_python.yaml +++ /dev/null @@ -1,11 +0,0 @@ -meta: -- desc: | - librbd python api tests -workload: - full_sequential: - - workunit: - branch: jewel - clients: - client.0: - - rbd/test_librbd_python.sh - - print: "**** done rbd/test_librbd_python.sh 2-workload" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/3-upgrade-sequence/upgrade-all.yaml b/src/ceph/qa/suites/upgrade/jewel-x/parallel/3-upgrade-sequence/upgrade-all.yaml deleted file mode 100644 index 6a0f829..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/3-upgrade-sequence/upgrade-all.yaml +++ /dev/null @@ -1,12 +0,0 @@ -meta: -- desc: | - upgrade the ceph cluster -upgrade-sequence: - sequential: - - ceph.restart: - daemons: [mon.a, mon.b, mon.c] - - ceph.restart: - daemons: [mds.a, osd.0, osd.1, osd.2, osd.3] - wait-for-healthy: false - wait-for-osds-up: true - - print: "**** done ceph.restart all" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/3-upgrade-sequence/upgrade-mon-osd-mds.yaml b/src/ceph/qa/suites/upgrade/jewel-x/parallel/3-upgrade-sequence/upgrade-mon-osd-mds.yaml deleted file mode 100644 index 2d74e9e..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/3-upgrade-sequence/upgrade-mon-osd-mds.yaml +++ /dev/null @@ -1,38 +0,0 @@ -meta: -- desc: | - upgrade the ceph cluster, - upgrate in two steps - step one ordering: mon.a, osd.0, osd.1, mds.a - step two ordering: mon.b, mon.c, osd.2, osd.3 - ceph expected to be healthy state after each step -upgrade-sequence: - sequential: - - ceph.restart: - daemons: [mon.a] - wait-for-healthy: true - - sleep: - duration: 60 - - ceph.restart: - daemons: [mon.b, mon.c] - wait-for-healthy: true - - sleep: - duration: 60 - - ceph.restart: - daemons: [osd.0, osd.1] - wait-for-healthy: true - - sleep: - duration: 60 - - ceph.restart: [mds.a] - - sleep: - duration: 60 - - print: "**** running mixed versions of osds and mons" - - exec: - mon.b: - - sudo ceph osd crush tunables jewel - - print: "**** done ceph osd crush tunables jewel" - - sleep: - duration: 60 - - ceph.restart: - daemons: [osd.2, osd.3] - wait-for-healthy: false - wait-for-osds-up: true diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/4-luminous.yaml b/src/ceph/qa/suites/upgrade/jewel-x/parallel/4-luminous.yaml deleted file mode 100644 index e57b377..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/4-luminous.yaml +++ /dev/null @@ -1,23 +0,0 @@ -# this is the same fragment as ../../../../releases/luminous.yaml -# but without line "ceph osd set-require-min-compat-client luminous" - -tasks: -- exec: - mgr.x: - - mkdir -p /var/lib/ceph/mgr/ceph-x - - ceph auth get-or-create-key mgr.x mon 'allow profile mgr' - - ceph auth export mgr.x > /var/lib/ceph/mgr/ceph-x/keyring -- ceph.restart: - daemons: [mgr.x] - wait-for-healthy: false -- exec: - osd.0: - - ceph osd require-osd-release luminous -- ceph.healthy: -overrides: - ceph: - conf: - mon: - mon warn on osd down out interval zero: false - log-whitelist: - - no active mgr diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/5-workload.yaml b/src/ceph/qa/suites/upgrade/jewel-x/parallel/5-workload.yaml deleted file mode 100644 index f7e9de4..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/5-workload.yaml +++ /dev/null @@ -1,11 +0,0 @@ -meta: -- desc: | - run basic import/export cli tests for rbd on not upgrated client.4 - (covers issue http://tracker.ceph.com/issues/21660) -tasks: - - workunit: - branch: jewel - clients: - client.4: - - rbd/import_export.sh - - print: "**** done rbd/import_export.sh 5-workload" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/6-luminous-with-mgr.yaml b/src/ceph/qa/suites/upgrade/jewel-x/parallel/6-luminous-with-mgr.yaml deleted file mode 120000 index 5c72153..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/6-luminous-with-mgr.yaml +++ /dev/null @@ -1 +0,0 @@ -../../../../releases/luminous-with-mgr.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/6.5-crush-compat.yaml b/src/ceph/qa/suites/upgrade/jewel-x/parallel/6.5-crush-compat.yaml deleted file mode 100644 index 20c0ffd..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/6.5-crush-compat.yaml +++ /dev/null @@ -1,8 +0,0 @@ -tasks: -- exec: - mon.a: - - ceph osd set-require-min-compat-client jewel - - ceph osd crush set-all-straw-buckets-to-straw2 - - ceph osd crush weight-set create-compat - - ceph osd crush weight-set reweight-compat osd.0 .9 - - ceph osd crush weight-set reweight-compat osd.1 1.2 diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/7-final-workload/+ b/src/ceph/qa/suites/upgrade/jewel-x/parallel/7-final-workload/+ deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/7-final-workload/+ +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/7-final-workload/blogbench.yaml b/src/ceph/qa/suites/upgrade/jewel-x/parallel/7-final-workload/blogbench.yaml deleted file mode 100644 index d73459e..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/7-final-workload/blogbench.yaml +++ /dev/null @@ -1,13 +0,0 @@ -meta: -- desc: | - run a cephfs stress test - mount ceph-fuse on client.3 before running workunit -tasks: -- sequential: - - ceph-fuse: [client.3] - - print: "**** done ceph-fuse 5-final-workload" - - workunit: - clients: - client.3: - - suites/blogbench.sh - - print: "**** done suites/blogbench.sh 7-final-workload" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/7-final-workload/rados-snaps-few-objects.yaml b/src/ceph/qa/suites/upgrade/jewel-x/parallel/7-final-workload/rados-snaps-few-objects.yaml deleted file mode 100644 index 7dd61c5..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/7-final-workload/rados-snaps-few-objects.yaml +++ /dev/null @@ -1,17 +0,0 @@ -meta: -- desc: | - randomized correctness test for rados operations on a replicated pool with snapshots -tasks: - - rados: - clients: [client.1] - ops: 4000 - objects: 50 - write_append_excl: false - op_weights: - read: 100 - write: 100 - delete: 50 - snap_create: 50 - snap_remove: 50 - rollback: 50 - - print: "**** done rados 7-final-workload" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/7-final-workload/rados_loadgenmix.yaml b/src/ceph/qa/suites/upgrade/jewel-x/parallel/7-final-workload/rados_loadgenmix.yaml deleted file mode 100644 index b218b92..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/7-final-workload/rados_loadgenmix.yaml +++ /dev/null @@ -1,9 +0,0 @@ -meta: -- desc: | - generate read/write load with rados objects ranging from 1 byte to 1MB -tasks: - - workunit: - clients: - client.1: - - rados/load-gen-mix.sh - - print: "**** done rados/load-gen-mix.sh 7-final-workload" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/7-final-workload/rados_mon_thrash.yaml b/src/ceph/qa/suites/upgrade/jewel-x/parallel/7-final-workload/rados_mon_thrash.yaml deleted file mode 100644 index c835a65..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/7-final-workload/rados_mon_thrash.yaml +++ /dev/null @@ -1,18 +0,0 @@ -meta: -- desc: | - librados C and C++ api tests -overrides: - ceph: - log-whitelist: - - reached quota -tasks: - - mon_thrash: - revive_delay: 20 - thrash_delay: 1 - - print: "**** done mon_thrash 4-final-workload" - - workunit: - branch: jewel - clients: - client.1: - - rados/test-upgrade-v11.0.0.sh - - print: "**** done rados/test-upgrade-v11.0.0.sh 7-final-workload" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/7-final-workload/rbd_cls.yaml b/src/ceph/qa/suites/upgrade/jewel-x/parallel/7-final-workload/rbd_cls.yaml deleted file mode 100644 index 46bbf76..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/7-final-workload/rbd_cls.yaml +++ /dev/null @@ -1,9 +0,0 @@ -meta: -- desc: | - rbd object class functional tests -tasks: - - workunit: - clients: - client.1: - - cls/test_cls_rbd.sh - - print: "**** done cls/test_cls_rbd.sh 7-final-workload" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/7-final-workload/rbd_import_export.yaml b/src/ceph/qa/suites/upgrade/jewel-x/parallel/7-final-workload/rbd_import_export.yaml deleted file mode 100644 index 5ae7491..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/7-final-workload/rbd_import_export.yaml +++ /dev/null @@ -1,11 +0,0 @@ -meta: -- desc: | - run basic import/export cli tests for rbd -tasks: - - workunit: - clients: - client.1: - - rbd/import_export.sh - env: - RBD_CREATE_ARGS: --new-format - - print: "**** done rbd/import_export.sh 7-final-workload" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/7-final-workload/rgw_swift.yaml b/src/ceph/qa/suites/upgrade/jewel-x/parallel/7-final-workload/rgw_swift.yaml deleted file mode 100644 index 780c4ad..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/7-final-workload/rgw_swift.yaml +++ /dev/null @@ -1,13 +0,0 @@ -meta: -- desc: | - swift api tests for rgw -overrides: - rgw: - frontend: civetweb -tasks: - - rgw: [client.1] - - print: "**** done rgw 7-final-workload" - - swift: - client.1: - rgw_server: client.1 - - print: "**** done swift 7-final-workload" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/8-jewel-workload.yaml b/src/ceph/qa/suites/upgrade/jewel-x/parallel/8-jewel-workload.yaml deleted file mode 120000 index 81df389..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/8-jewel-workload.yaml +++ /dev/null @@ -1 +0,0 @@ -5-workload.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/jewel-x/parallel/distros b/src/ceph/qa/suites/upgrade/jewel-x/parallel/distros deleted file mode 120000 index ca99fee..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/parallel/distros +++ /dev/null @@ -1 +0,0 @@ -../../../../distros/supported/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/jewel-x/point-to-point-x/% b/src/ceph/qa/suites/upgrade/jewel-x/point-to-point-x/% deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/point-to-point-x/% +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/jewel-x/point-to-point-x/distros/centos_7.3.yaml b/src/ceph/qa/suites/upgrade/jewel-x/point-to-point-x/distros/centos_7.3.yaml deleted file mode 120000 index c79327b..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/point-to-point-x/distros/centos_7.3.yaml +++ /dev/null @@ -1 +0,0 @@ -../../../../../distros/all/centos_7.3.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/jewel-x/point-to-point-x/distros/ubuntu_14.04.yaml b/src/ceph/qa/suites/upgrade/jewel-x/point-to-point-x/distros/ubuntu_14.04.yaml deleted file mode 120000 index 6237042..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/point-to-point-x/distros/ubuntu_14.04.yaml +++ /dev/null @@ -1 +0,0 @@ -../../../../../distros/all/ubuntu_14.04.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/jewel-x/point-to-point-x/point-to-point-upgrade.yaml b/src/ceph/qa/suites/upgrade/jewel-x/point-to-point-x/point-to-point-upgrade.yaml deleted file mode 100644 index d68c258..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/point-to-point-x/point-to-point-upgrade.yaml +++ /dev/null @@ -1,236 +0,0 @@ -meta: -- desc: | - Run ceph on two nodes, using one of them as a client, - with a separate client-only node. - Use xfs beneath the osds. - install ceph/jewel v10.2.0 point version - run workload and upgrade-sequence in parallel - install ceph/jewel latest version - run workload and upgrade-sequence in parallel - install ceph/-x version (jewel or kraken) - run workload and upgrade-sequence in parallel -overrides: - ceph: - log-whitelist: - - reached quota - - scrub - - osd_map_max_advance - - wrongly marked - - overall HEALTH_ - - \(MGR_DOWN\) - - \(OSD_ - - \(PG_ - - \(CACHE_ - fs: xfs - conf: - global: - mon warn on pool no app: false - mon: - mon debug unsafe allow tier with nonempty snaps: true - osd: - osd map max advance: 1000 - osd map cache size: 1100 -roles: -- - mon.a - - mds.a - - osd.0 - - osd.1 - - osd.2 - - mgr.x -- - mon.b - - mon.c - - osd.3 - - osd.4 - - osd.5 - - client.0 -- - client.1 -openstack: -- volumes: # attached to each instance - count: 3 - size: 30 # GB -tasks: -- print: "**** v10.2.0 about to install" -- install: - tag: v10.2.0 - exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev', 'librgw2'] -- print: "**** done v10.2.0 install" -- ceph: - fs: xfs - skip_mgr_daemons: true - add_osds_to_crush: true -- print: "**** done ceph xfs" -- sequential: - - workload -- print: "**** done workload v10.2.0" -- install.upgrade: - exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] - mon.a: - branch: jewel - mon.b: - branch: jewel - # Note that client.a IS NOT upgraded at this point - #client.1: - #branch: jewel -- parallel: - - workload_jewel - - upgrade-sequence_jewel -- print: "**** done parallel jewel branch" -- install.upgrade: - exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] - client.1: - branch: jewel -- print: "**** done branch: jewel install.upgrade on client.1" -- install.upgrade: - mon.a: - mon.b: -- print: "**** done branch: -x install.upgrade on mon.a and mon.b" -- parallel: - - workload_x - - upgrade-sequence_x -- print: "**** done parallel -x branch" -- exec: - osd.0: - - ceph osd set-require-min-compat-client luminous -# Run librados tests on the -x upgraded cluster -- install.upgrade: - client.1: -- workunit: - branch: jewel - clients: - client.1: - - rados/test-upgrade-v11.0.0.sh - - cls -- print: "**** done final test on -x cluster" -####################### -workload: - sequential: - - workunit: - clients: - client.0: - - suites/blogbench.sh -workload_jewel: - full_sequential: - - workunit: - branch: jewel - clients: - client.1: - - rados/test.sh - - cls - env: - CLS_RBD_GTEST_FILTER: '*:-TestClsRbd.mirror_image' - - print: "**** done rados/test.sh & cls workload_jewel" - - sequential: - - rgw: [client.0] - - print: "**** done rgw workload_jewel" - - s3tests: - client.0: - force-branch: ceph-jewel - rgw_server: client.0 - scan_for_encryption_keys: false - - print: "**** done s3tests workload_jewel" -upgrade-sequence_jewel: - sequential: - - print: "**** done branch: jewel install.upgrade" - - ceph.restart: [mds.a] - - sleep: - duration: 60 - - ceph.restart: [osd.0] - - sleep: - duration: 30 - - ceph.restart: [osd.1] - - sleep: - duration: 30 - - ceph.restart: [osd.2] - - sleep: - duration: 30 - - ceph.restart: [osd.3] - - sleep: - duration: 30 - - ceph.restart: [osd.4] - - sleep: - duration: 30 - - ceph.restart: [osd.5] - - sleep: - duration: 60 - - ceph.restart: [mon.a] - - sleep: - duration: 60 - - ceph.restart: [mon.b] - - sleep: - duration: 60 - - ceph.restart: [mon.c] - - sleep: - duration: 60 - - print: "**** done ceph.restart all jewel branch mds/osd/mon" -workload_x: - sequential: - - workunit: - branch: jewel - clients: - client.1: - - rados/test-upgrade-v11.0.0-noec.sh - - cls - env: - CLS_RBD_GTEST_FILTER: '*:-TestClsRbd.mirror_image' - - print: "**** done rados/test-upgrade-v11.0.0.sh & cls workload_x NOT upgraded client" - - workunit: - branch: jewel - clients: - client.0: - - rados/test-upgrade-v11.0.0-noec.sh - - cls - - print: "**** done rados/test-upgrade-v11.0.0.sh & cls workload_x upgraded client" - - rgw: [client.1] - - print: "**** done rgw workload_x" - - s3tests: - client.1: - force-branch: ceph-jewel - rgw_server: client.1 - scan_for_encryption_keys: false - - print: "**** done s3tests workload_x" -upgrade-sequence_x: - sequential: - - ceph.restart: [mds.a] - - sleep: - duration: 60 - - ceph.restart: [mon.a] - - sleep: - duration: 60 - - ceph.restart: [mon.b] - - sleep: - duration: 60 - - ceph.restart: [mon.c] - - sleep: - duration: 60 - - ceph.restart: [osd.0] - - sleep: - duration: 30 - - ceph.restart: [osd.1] - - sleep: - duration: 30 - - ceph.restart: [osd.2] - - sleep: - duration: 30 - - ceph.restart: [osd.3] - - sleep: - duration: 30 - - ceph.restart: [osd.4] - - sleep: - duration: 30 - - ceph.restart: - daemons: [osd.5] - wait-for-healthy: false - wait-for-up-osds: true - - exec: - mgr.x: - - mkdir -p /var/lib/ceph/mgr/ceph-x - - ceph auth get-or-create-key mgr.x mon 'allow profile mgr' - - ceph auth export mgr.x > /var/lib/ceph/mgr/ceph-x/keyring - - ceph.restart: - daemons: [mgr.x] - wait-for-healthy: false - - exec: - osd.0: - - ceph osd require-osd-release luminous - - ceph.healthy: - - print: "**** done ceph.restart all -x branch mds/osd/mon" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/% b/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/% deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/% +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/0-cluster b/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/0-cluster deleted file mode 120000 index 3580937..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/0-cluster +++ /dev/null @@ -1 +0,0 @@ -../stress-split/0-cluster/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/1-jewel-install b/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/1-jewel-install deleted file mode 120000 index 3e7cbc3..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/1-jewel-install +++ /dev/null @@ -1 +0,0 @@ -../stress-split/1-jewel-install/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/1.5-final-scrub.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/1.5-final-scrub.yaml deleted file mode 120000 index 522db1b..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/1.5-final-scrub.yaml +++ /dev/null @@ -1 +0,0 @@ -../parallel/1.5-final-scrub.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/2-partial-upgrade b/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/2-partial-upgrade deleted file mode 120000 index ab35fc1..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/2-partial-upgrade +++ /dev/null @@ -1 +0,0 @@ -../stress-split/2-partial-upgrade/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/3-thrash/default.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/3-thrash/default.yaml deleted file mode 100644 index edae7b3..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/3-thrash/default.yaml +++ /dev/null @@ -1,25 +0,0 @@ -meta: -- desc: | - randomly kill and revive osd - small chance to increase the number of pgs -overrides: - ceph: - log-whitelist: - - but it is still running - - wrongly marked me down - - objects unfound and apparently lost - - log bound mismatch -tasks: -- parallel: - - stress-tasks -stress-tasks: -- thrashosds: - timeout: 1200 - chance_pgnum_grow: 1 - chance_pgpnum_fix: 1 - min_in: 4 - chance_thrash_cluster_full: 0 - chance_thrash_pg_upmap: 0 - chance_thrash_pg_upmap_items: 0 - chance_force_recovery: 0 -- print: "**** done thrashosds 3-thrash" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/4-workload/ec-rados-default.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/4-workload/ec-rados-default.yaml deleted file mode 100644 index c89551e..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/4-workload/ec-rados-default.yaml +++ /dev/null @@ -1,22 +0,0 @@ -meta: -- desc: | - randomized correctness test for rados operations on an erasure coded pool -stress-tasks: - - rados: - clients: [client.0] - ops: 4000 - objects: 50 - ec_pool: true - write_append_excl: false - op_weights: - read: 100 - write: 0 - append: 100 - delete: 50 - snap_create: 50 - snap_remove: 50 - rollback: 50 - copy_from: 50 - setattr: 25 - rmattr: 25 - - print: "**** done rados ec task" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/5-finish-upgrade.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/5-finish-upgrade.yaml deleted file mode 120000 index a66a7dc..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/5-finish-upgrade.yaml +++ /dev/null @@ -1 +0,0 @@ -../stress-split/5-finish-upgrade.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/6-luminous.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/6-luminous.yaml deleted file mode 120000 index 2b99d5c..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/6-luminous.yaml +++ /dev/null @@ -1 +0,0 @@ -../stress-split/6-luminous.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/7-final-workload/ec-rados-plugin=jerasure-k=3-m=1.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/7-final-workload/ec-rados-plugin=jerasure-k=3-m=1.yaml deleted file mode 100644 index a82f11b..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/7-final-workload/ec-rados-plugin=jerasure-k=3-m=1.yaml +++ /dev/null @@ -1,35 +0,0 @@ -# -# k=3 implies a stripe_width of 1376*3 = 4128 which is different from -# the default value of 4096 It is also not a multiple of 1024*1024 and -# creates situations where rounding rules during recovery becomes -# necessary. -# -meta: -- desc: | - randomized correctness test for rados operations on an erasure coded pool - using the jerasure plugin with k=3 and m=1 -tasks: -- rados: - clients: [client.0] - ops: 4000 - objects: 50 - ec_pool: true - write_append_excl: false - erasure_code_profile: - name: jerasure31profile - plugin: jerasure - k: 3 - m: 1 - technique: reed_sol_van - crush-failure-domain: osd - op_weights: - read: 100 - write: 0 - append: 100 - delete: 50 - snap_create: 50 - snap_remove: 50 - rollback: 50 - copy_from: 50 - setattr: 25 - rmattr: 25 diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/distros b/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/distros deleted file mode 120000 index ca99fee..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/distros +++ /dev/null @@ -1 +0,0 @@ -../../../../distros/supported/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/thrashosds-health.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/thrashosds-health.yaml deleted file mode 120000 index e0426db..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split-erasure-code/thrashosds-health.yaml +++ /dev/null @@ -1 +0,0 @@ -../../../../tasks/thrashosds-health.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/% b/src/ceph/qa/suites/upgrade/jewel-x/stress-split/% deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/% +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/0-cluster/+ b/src/ceph/qa/suites/upgrade/jewel-x/stress-split/0-cluster/+ deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/0-cluster/+ +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/0-cluster/openstack.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split/0-cluster/openstack.yaml deleted file mode 100644 index a0d5c20..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/0-cluster/openstack.yaml +++ /dev/null @@ -1,6 +0,0 @@ -openstack: - - machine: - disk: 100 # GB - - volumes: # attached to each instance - count: 3 - size: 30 # GB diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/0-cluster/start.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split/0-cluster/start.yaml deleted file mode 100644 index 4f40219..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/0-cluster/start.yaml +++ /dev/null @@ -1,20 +0,0 @@ -meta: -- desc: | - Run ceph on two nodes, - with a separate client-only node. - Use xfs beneath the osds. -overrides: - ceph: - fs: xfs -roles: -- - mon.a - - mon.b - - mon.c - - mgr.x - - osd.0 - - osd.1 - - osd.2 -- - osd.3 - - osd.4 - - osd.5 -- - client.0 diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/1-jewel-install/jewel.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split/1-jewel-install/jewel.yaml deleted file mode 100644 index 31ca3e5..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/1-jewel-install/jewel.yaml +++ /dev/null @@ -1,13 +0,0 @@ -meta: -- desc: install ceph/jewel latest -tasks: -- install: - branch: jewel - exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] -- print: "**** done install jewel" -- ceph: - skip_mgr_daemons: true - add_osds_to_crush: true - log-whitelist: - - required past_interval bounds are empty -- print: "**** done ceph" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/1.5-final-scrub.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split/1.5-final-scrub.yaml deleted file mode 120000 index 522db1b..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/1.5-final-scrub.yaml +++ /dev/null @@ -1 +0,0 @@ -../parallel/1.5-final-scrub.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/2-partial-upgrade/firsthalf.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split/2-partial-upgrade/firsthalf.yaml deleted file mode 100644 index 442dcf1..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/2-partial-upgrade/firsthalf.yaml +++ /dev/null @@ -1,12 +0,0 @@ -meta: -- desc: | - install upgrade ceph/-x on one node only - 1st half - restart : osd.0,1,2 -tasks: -- install.upgrade: - osd.0: -- print: "**** done install.upgrade osd.0" -- ceph.restart: - daemons: [mon.a,mon.b,mon.c,osd.0, osd.1, osd.2] -- print: "**** done ceph.restart 1st half" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/3-thrash/default.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split/3-thrash/default.yaml deleted file mode 100644 index b3fddef..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/3-thrash/default.yaml +++ /dev/null @@ -1,25 +0,0 @@ -meta: -- desc: | - randomly kill and revive osd - small chance to increase the number of pgs -overrides: - ceph: - log-whitelist: - - but it is still running - - wrongly marked me down - - objects unfound and apparently lost - - log bound mismatch -tasks: -- parallel: - - stress-tasks -stress-tasks: -- thrashosds: - timeout: 1200 - chance_pgnum_grow: 1 - chance_pgpnum_fix: 1 - chance_thrash_cluster_full: 0 - chance_thrash_pg_upmap: 0 - chance_thrash_pg_upmap_items: 0 - disable_objectstore_tool_tests: true - chance_force_recovery: 0 -- print: "**** done thrashosds 3-thrash" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/4-workload/+ b/src/ceph/qa/suites/upgrade/jewel-x/stress-split/4-workload/+ deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/4-workload/+ +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/4-workload/radosbench.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split/4-workload/radosbench.yaml deleted file mode 100644 index 626ae8e..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/4-workload/radosbench.yaml +++ /dev/null @@ -1,40 +0,0 @@ -meta: -- desc: | - run randomized correctness test for rados operations - generate write load with rados bench -stress-tasks: -- full_sequential: - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 -- print: "**** done radosbench 7-workload" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/4-workload/rbd-cls.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split/4-workload/rbd-cls.yaml deleted file mode 100644 index 92779bc..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/4-workload/rbd-cls.yaml +++ /dev/null @@ -1,10 +0,0 @@ -meta: -- desc: | - run basic cls tests for rbd -stress-tasks: -- workunit: - branch: jewel - clients: - client.0: - - cls/test_cls_rbd.sh -- print: "**** done cls/test_cls_rbd.sh 5-workload" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/4-workload/rbd-import-export.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split/4-workload/rbd-import-export.yaml deleted file mode 100644 index 693154d..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/4-workload/rbd-import-export.yaml +++ /dev/null @@ -1,12 +0,0 @@ -meta: -- desc: | - run basic import/export cli tests for rbd -stress-tasks: -- workunit: - branch: jewel - clients: - client.0: - - rbd/import_export.sh - env: - RBD_CREATE_ARGS: --new-format -- print: "**** done rbd/import_export.sh 5-workload" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/4-workload/rbd_api.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split/4-workload/rbd_api.yaml deleted file mode 100644 index 64c0e33..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/4-workload/rbd_api.yaml +++ /dev/null @@ -1,10 +0,0 @@ -meta: -- desc: | - librbd C and C++ api tests -stress-tasks: -- workunit: - branch: jewel - clients: - client.0: - - rbd/test_librbd.sh -- print: "**** done rbd/test_librbd.sh 7-workload" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/4-workload/readwrite.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split/4-workload/readwrite.yaml deleted file mode 100644 index 41e34d6..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/4-workload/readwrite.yaml +++ /dev/null @@ -1,16 +0,0 @@ -meta: -- desc: | - randomized correctness test for rados operations on a replicated pool, - using only reads, writes, and deletes -stress-tasks: -- full_sequential: - - rados: - clients: [client.0] - ops: 4000 - objects: 500 - write_append_excl: false - op_weights: - read: 45 - write: 45 - delete: 10 -- print: "**** done rados/readwrite 5-workload" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/4-workload/snaps-few-objects.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split/4-workload/snaps-few-objects.yaml deleted file mode 100644 index f56d0de..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/4-workload/snaps-few-objects.yaml +++ /dev/null @@ -1,18 +0,0 @@ -meta: -- desc: | - randomized correctness test for rados operations on a replicated pool with snapshot operations -stress-tasks: -- full_sequential: - - rados: - clients: [client.0] - ops: 4000 - objects: 50 - write_append_excl: false - op_weights: - read: 100 - write: 100 - delete: 50 - snap_create: 50 - snap_remove: 50 - rollback: 50 -- print: "**** done rados/snaps-few-objects 5-workload" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/5-finish-upgrade.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split/5-finish-upgrade.yaml deleted file mode 100644 index 1d528cd..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/5-finish-upgrade.yaml +++ /dev/null @@ -1,9 +0,0 @@ -tasks: -- install.upgrade: - osd.3: - client.0: -- ceph.restart: - daemons: [osd.3, osd.4, osd.5] - wait-for-healthy: false - wait-for-osds-up: true - diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/6-luminous.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split/6-luminous.yaml deleted file mode 120000 index 5283ac7..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/6-luminous.yaml +++ /dev/null @@ -1 +0,0 @@ -../../../../releases/luminous.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/6.5-crush-compat.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split/6.5-crush-compat.yaml deleted file mode 120000 index 02263d1..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/6.5-crush-compat.yaml +++ /dev/null @@ -1 +0,0 @@ -../parallel/6.5-crush-compat.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/7-final-workload/+ b/src/ceph/qa/suites/upgrade/jewel-x/stress-split/7-final-workload/+ deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/7-final-workload/+ +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/7-final-workload/rbd-python.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split/7-final-workload/rbd-python.yaml deleted file mode 100644 index 56ba21d..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/7-final-workload/rbd-python.yaml +++ /dev/null @@ -1,9 +0,0 @@ -meta: -- desc: | - librbd python api tests -tasks: -- workunit: - clients: - client.0: - - rbd/test_librbd_python.sh -- print: "**** done rbd/test_librbd_python.sh 9-workload" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/7-final-workload/rgw-swift.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split/7-final-workload/rgw-swift.yaml deleted file mode 100644 index 76e5d6f..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/7-final-workload/rgw-swift.yaml +++ /dev/null @@ -1,11 +0,0 @@ -meta: -- desc: | - swift api tests for rgw -tasks: -- rgw: - client.0: -- print: "**** done rgw 9-workload" -- swift: - client.0: - rgw_server: client.0 -- print: "**** done swift 9-workload" diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/7-final-workload/snaps-many-objects.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split/7-final-workload/snaps-many-objects.yaml deleted file mode 100644 index 805bf97..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/7-final-workload/snaps-many-objects.yaml +++ /dev/null @@ -1,16 +0,0 @@ -meta: -- desc: | - randomized correctness test for rados operations on a replicated pool with snapshot operations -tasks: -- rados: - clients: [client.0] - ops: 4000 - objects: 500 - write_append_excl: false - op_weights: - read: 100 - write: 100 - delete: 50 - snap_create: 50 - snap_remove: 50 - rollback: 50 diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/distros b/src/ceph/qa/suites/upgrade/jewel-x/stress-split/distros deleted file mode 120000 index ca99fee..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/distros +++ /dev/null @@ -1 +0,0 @@ -../../../../distros/supported/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/thrashosds-health.yaml b/src/ceph/qa/suites/upgrade/jewel-x/stress-split/thrashosds-health.yaml deleted file mode 120000 index e0426db..0000000 --- a/src/ceph/qa/suites/upgrade/jewel-x/stress-split/thrashosds-health.yaml +++ /dev/null @@ -1 +0,0 @@ -../../../../tasks/thrashosds-health.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/kraken-x/ceph-deploy/kraken-luminous.yaml b/src/ceph/qa/suites/upgrade/kraken-x/ceph-deploy/kraken-luminous.yaml deleted file mode 100644 index 4a55362..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/ceph-deploy/kraken-luminous.yaml +++ /dev/null @@ -1,61 +0,0 @@ -meta: -- desc: | - Setup 4 node ceph cluster using ceph-deploy, use latest - stable kraken as initial release, upgrade to luminous and - also setup mgr nodes along after upgrade, check for - cluster to reach healthy state, After upgrade run kernel tar/untar - task and systemd task. This test will detect any - ceph upgrade issue and systemd issues. -overrides: - ceph-deploy: - fs: xfs - conf: - global: - mon pg warn min per osd: 2 - osd: - osd pool default size: 2 - osd objectstore: filestore - osd sloppy crc: true - client: - rbd default features: 5 -roles: -- - mon.a - - mds.a - - osd.0 - - osd.1 - - osd.2 - - mgr.x -- - mon.b - - mgr.y -- - mon.c - - osd.3 - - osd.4 - - osd.5 -- - osd.6 - - osd.7 - - osd.8 - - client.0 -tasks: -- ssh-keys: -- ceph-deploy: - branch: - stable: kraken - skip-mgr: True -- ceph-deploy.upgrade: - branch: - dev: luminous - setup-mgr-node: True - check-for-healthy: True - roles: - - mon.a - - mon.b - - mon.c -- workunit: - clients: - all: - - kernel_untar_build.sh -- systemd: -- workunit: - clients: - all: - - rados/load-gen-mix.sh diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/% b/src/ceph/qa/suites/upgrade/kraken-x/parallel/% deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/% +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/0-cluster/+ b/src/ceph/qa/suites/upgrade/kraken-x/parallel/0-cluster/+ deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/0-cluster/+ +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/0-cluster/openstack.yaml b/src/ceph/qa/suites/upgrade/kraken-x/parallel/0-cluster/openstack.yaml deleted file mode 100644 index f4d1349..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/0-cluster/openstack.yaml +++ /dev/null @@ -1,4 +0,0 @@ -openstack: - - volumes: # attached to each instance - count: 3 - size: 30 # GB diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/0-cluster/start.yaml b/src/ceph/qa/suites/upgrade/kraken-x/parallel/0-cluster/start.yaml deleted file mode 100644 index f5a883a..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/0-cluster/start.yaml +++ /dev/null @@ -1,33 +0,0 @@ -meta: -- desc: | - Run ceph on two nodes, - with a separate client 0,1,2 third node. - Use xfs beneath the osds. - CephFS tests running on client 2,3 -roles: -- - mon.a - - mgr.x - - mds.a - - osd.0 - - osd.1 -- - mon.b - - mon.c - - osd.2 - - osd.3 -- - client.0 - - client.1 - - client.2 - - client.3 -- - client.4 -overrides: - ceph: - log-whitelist: - - scrub mismatch - - ScrubResult - - wrongly marked - - (POOL_APP_NOT_ENABLED) - - overall HEALTH_ - conf: - global: - enable experimental unrecoverable data corrupting features: "*" - fs: xfs diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/1-kraken-install/kraken.yaml b/src/ceph/qa/suites/upgrade/kraken-x/parallel/1-kraken-install/kraken.yaml deleted file mode 100644 index de0893c..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/1-kraken-install/kraken.yaml +++ /dev/null @@ -1,39 +0,0 @@ -meta: -- desc: | - install ceph/kraken latest - run workload and upgrade-sequence in parallel - upgrade the client node -tasks: -- install: - branch: kraken -- print: "**** done installing kraken" -- ceph: - log-whitelist: - - overall HEALTH_ - - \(FS_ - - \(MDS_ - - \(OSD_ - - \(MON_DOWN\) - - \(CACHE_POOL_ - - \(POOL_ - - \(MGR_DOWN\) - - \(PG_ - - \(SMALLER_PGP_NUM\) - - Monitor daemon marked osd - - Behind on trimming - - Manager daemon - conf: - global: - mon warn on pool no app: false -- print: "**** done ceph" -- install.upgrade: - mon.a: - mon.b: -- print: "**** done install.upgrade both hosts" -- parallel: - - workload - - upgrade-sequence -- print: "**** done parallel" -- install.upgrade: - client.0: -- print: "**** done install.upgrade on client.0" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/2-workload/+ b/src/ceph/qa/suites/upgrade/kraken-x/parallel/2-workload/+ deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/2-workload/+ +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/2-workload/blogbench.yaml b/src/ceph/qa/suites/upgrade/kraken-x/parallel/2-workload/blogbench.yaml deleted file mode 100644 index 021fcc6..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/2-workload/blogbench.yaml +++ /dev/null @@ -1,14 +0,0 @@ -meta: -- desc: | - run a cephfs stress test - mount ceph-fuse on client.2 before running workunit -workload: - full_sequential: - - sequential: - - ceph-fuse: - - print: "**** done ceph-fuse 2-workload" - - workunit: - clients: - client.2: - - suites/blogbench.sh - - print: "**** done suites/blogbench.sh 2-workload" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/2-workload/ec-rados-default.yaml b/src/ceph/qa/suites/upgrade/kraken-x/parallel/2-workload/ec-rados-default.yaml deleted file mode 100644 index 5c5a958..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/2-workload/ec-rados-default.yaml +++ /dev/null @@ -1,24 +0,0 @@ -meta: -- desc: | - run run randomized correctness test for rados operations - on an erasure-coded pool -workload: - full_sequential: - - rados: - clients: [client.0] - ops: 4000 - objects: 50 - ec_pool: true - write_append_excl: false - op_weights: - read: 100 - write: 0 - append: 100 - delete: 50 - snap_create: 50 - snap_remove: 50 - rollback: 50 - copy_from: 50 - setattr: 25 - rmattr: 25 - - print: "**** done rados ec task" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/2-workload/rados_api.yaml b/src/ceph/qa/suites/upgrade/kraken-x/parallel/2-workload/rados_api.yaml deleted file mode 100644 index 893beec..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/2-workload/rados_api.yaml +++ /dev/null @@ -1,11 +0,0 @@ -meta: -- desc: | - object class functional tests -workload: - full_sequential: - - workunit: - branch: kraken - clients: - client.0: - - cls - - print: "**** done cls 2-workload" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/2-workload/rados_loadgenbig.yaml b/src/ceph/qa/suites/upgrade/kraken-x/parallel/2-workload/rados_loadgenbig.yaml deleted file mode 100644 index 8befdd4..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/2-workload/rados_loadgenbig.yaml +++ /dev/null @@ -1,11 +0,0 @@ -meta: -- desc: | - generate read/write load with rados objects ranging from 1MB to 25MB -workload: - full_sequential: - - workunit: - branch: kraken - clients: - client.0: - - rados/load-gen-big.sh - - print: "**** done rados/load-gen-big.sh 2-workload" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/2-workload/test_rbd_api.yaml b/src/ceph/qa/suites/upgrade/kraken-x/parallel/2-workload/test_rbd_api.yaml deleted file mode 100644 index 10f4b05..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/2-workload/test_rbd_api.yaml +++ /dev/null @@ -1,11 +0,0 @@ -meta: -- desc: | - librbd C and C++ api tests -workload: - full_sequential: - - workunit: - branch: kraken - clients: - client.0: - - rbd/test_librbd.sh - - print: "**** done rbd/test_librbd.sh 2-workload" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/2-workload/test_rbd_python.yaml b/src/ceph/qa/suites/upgrade/kraken-x/parallel/2-workload/test_rbd_python.yaml deleted file mode 100644 index 23e653d..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/2-workload/test_rbd_python.yaml +++ /dev/null @@ -1,11 +0,0 @@ -meta: -- desc: | - librbd python api tests -workload: - full_sequential: - - workunit: - branch: kraken - clients: - client.0: - - rbd/test_librbd_python.sh - - print: "**** done rbd/test_librbd_python.sh 2-workload" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/3-upgrade-sequence/upgrade-all.yaml b/src/ceph/qa/suites/upgrade/kraken-x/parallel/3-upgrade-sequence/upgrade-all.yaml deleted file mode 100644 index cff3a68..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/3-upgrade-sequence/upgrade-all.yaml +++ /dev/null @@ -1,16 +0,0 @@ -meta: -- desc: | - upgrade the ceph cluster -upgrade-sequence: - sequential: - - ceph.restart: - daemons: [mon.a, mon.b, mon.c, mgr.x] - - ceph.restart: - daemons: [osd.0, osd.1, osd.2, osd.3] - wait-for-healthy: false - wait-for-osds-up: true - - ceph.restart: - daemons: [mds.a] - wait-for-healthy: false - wait-for-osds-up: true - - print: "**** done ceph.restart all" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/3-upgrade-sequence/upgrade-mon-osd-mds.yaml b/src/ceph/qa/suites/upgrade/kraken-x/parallel/3-upgrade-sequence/upgrade-mon-osd-mds.yaml deleted file mode 100644 index f197de6..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/3-upgrade-sequence/upgrade-mon-osd-mds.yaml +++ /dev/null @@ -1,35 +0,0 @@ -meta: -- desc: | - upgrade the ceph cluster, - upgrate in two steps - step one ordering: mon.a, osd.0, osd.1, mds.a - step two ordering: mon.b, mon.c, osd.2, osd.3 - ceph expected to be healthy state after each step -upgrade-sequence: - sequential: - - ceph.restart: - daemons: [mon.a] - wait-for-healthy: true - - sleep: - duration: 60 - - ceph.restart: - daemons: [mon.b, mon.c, mgr.x] - wait-for-healthy: true - - sleep: - duration: 60 - - ceph.restart: - daemons: [osd.0, osd.1] - wait-for-healthy: true - - sleep: - duration: 60 - - ceph.restart: [mds.a] - - sleep: - duration: 60 - - sleep: - duration: 60 - - ceph.restart: - daemons: [osd.2, osd.3] - wait-for-healthy: false - wait-for-osds-up: true - - sleep: - duration: 60 diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/4-luminous.yaml b/src/ceph/qa/suites/upgrade/kraken-x/parallel/4-luminous.yaml deleted file mode 100644 index 80c2b9d..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/4-luminous.yaml +++ /dev/null @@ -1,4 +0,0 @@ -tasks: -- exec: - osd.0: - - ceph osd require-osd-release luminous diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/5-workload.yaml b/src/ceph/qa/suites/upgrade/kraken-x/parallel/5-workload.yaml deleted file mode 100644 index 851c5c8..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/5-workload.yaml +++ /dev/null @@ -1,11 +0,0 @@ -meta: -- desc: | - run basic import/export cli tests for rbd on not upgrated client.4 - (covers issue http://tracker.ceph.com/issues/21660) -tasks: - - workunit: - branch: kraken - clients: - client.4: - - rbd/import_export.sh - - print: "**** done rbd/import_export.sh 5-workload" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/6-luminous-with-mgr.yaml b/src/ceph/qa/suites/upgrade/kraken-x/parallel/6-luminous-with-mgr.yaml deleted file mode 120000 index 5c72153..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/6-luminous-with-mgr.yaml +++ /dev/null @@ -1 +0,0 @@ -../../../../releases/luminous-with-mgr.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/7-final-workload/+ b/src/ceph/qa/suites/upgrade/kraken-x/parallel/7-final-workload/+ deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/7-final-workload/+ +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/7-final-workload/blogbench.yaml b/src/ceph/qa/suites/upgrade/kraken-x/parallel/7-final-workload/blogbench.yaml deleted file mode 100644 index d2629c0..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/7-final-workload/blogbench.yaml +++ /dev/null @@ -1,13 +0,0 @@ -meta: -- desc: | - run a cephfs stress test - mount ceph-fuse on client.3 before running workunit -tasks: -- sequential: - - ceph-fuse: - - print: "**** done ceph-fuse 5-final-workload" - - workunit: - clients: - client.3: - - suites/blogbench.sh - - print: "**** done suites/blogbench.sh 5-final-workload" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/7-final-workload/rados-snaps-few-objects.yaml b/src/ceph/qa/suites/upgrade/kraken-x/parallel/7-final-workload/rados-snaps-few-objects.yaml deleted file mode 100644 index d8b3dcb..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/7-final-workload/rados-snaps-few-objects.yaml +++ /dev/null @@ -1,17 +0,0 @@ -meta: -- desc: | - randomized correctness test for rados operations on a replicated pool with snapshots -tasks: - - rados: - clients: [client.1] - ops: 4000 - objects: 50 - write_append_excl: false - op_weights: - read: 100 - write: 100 - delete: 50 - snap_create: 50 - snap_remove: 50 - rollback: 50 - - print: "**** done rados 4-final-workload" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/7-final-workload/rados_loadgenmix.yaml b/src/ceph/qa/suites/upgrade/kraken-x/parallel/7-final-workload/rados_loadgenmix.yaml deleted file mode 100644 index 922a9da..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/7-final-workload/rados_loadgenmix.yaml +++ /dev/null @@ -1,9 +0,0 @@ -meta: -- desc: | - generate read/write load with rados objects ranging from 1 byte to 1MB -tasks: - - workunit: - clients: - client.1: - - rados/load-gen-mix.sh - - print: "**** done rados/load-gen-mix.sh 4-final-workload" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/7-final-workload/rados_mon_thrash.yaml b/src/ceph/qa/suites/upgrade/kraken-x/parallel/7-final-workload/rados_mon_thrash.yaml deleted file mode 100644 index ab6276e..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/7-final-workload/rados_mon_thrash.yaml +++ /dev/null @@ -1,18 +0,0 @@ -meta: -- desc: | - librados C and C++ api tests -overrides: - ceph: - log-whitelist: - - reached quota -tasks: - - mon_thrash: - revive_delay: 20 - thrash_delay: 1 - - print: "**** done mon_thrash 4-final-workload" - - workunit: - branch: kraken - clients: - client.1: - - rados/test.sh - - print: "**** done rados/test.sh 4-final-workload" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/7-final-workload/rbd_cls.yaml b/src/ceph/qa/suites/upgrade/kraken-x/parallel/7-final-workload/rbd_cls.yaml deleted file mode 100644 index aaf0a37..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/7-final-workload/rbd_cls.yaml +++ /dev/null @@ -1,9 +0,0 @@ -meta: -- desc: | - rbd object class functional tests -tasks: - - workunit: - clients: - client.1: - - cls/test_cls_rbd.sh - - print: "**** done cls/test_cls_rbd.sh 4-final-workload" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/7-final-workload/rbd_import_export.yaml b/src/ceph/qa/suites/upgrade/kraken-x/parallel/7-final-workload/rbd_import_export.yaml deleted file mode 100644 index 46e1355..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/7-final-workload/rbd_import_export.yaml +++ /dev/null @@ -1,11 +0,0 @@ -meta: -- desc: | - run basic import/export cli tests for rbd -tasks: - - workunit: - clients: - client.1: - - rbd/import_export.sh - env: - RBD_CREATE_ARGS: --new-format - - print: "**** done rbd/import_export.sh 4-final-workload" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/7-final-workload/rgw_swift.yaml b/src/ceph/qa/suites/upgrade/kraken-x/parallel/7-final-workload/rgw_swift.yaml deleted file mode 100644 index 7a7659f..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/7-final-workload/rgw_swift.yaml +++ /dev/null @@ -1,13 +0,0 @@ -meta: -- desc: | - swift api tests for rgw -overrides: - rgw: - frontend: civetweb -tasks: - - rgw: [client.1] - - print: "**** done rgw 4-final-workload" - - swift: - client.1: - rgw_server: client.1 - - print: "**** done swift 4-final-workload" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/distros b/src/ceph/qa/suites/upgrade/kraken-x/parallel/distros deleted file mode 120000 index ca99fee..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/distros +++ /dev/null @@ -1 +0,0 @@ -../../../../distros/supported/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/kraken-x/parallel/objectstore b/src/ceph/qa/suites/upgrade/kraken-x/parallel/objectstore deleted file mode 120000 index 016cbf9..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/parallel/objectstore +++ /dev/null @@ -1 +0,0 @@ -../stress-split/objectstore/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/% b/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/% deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/% +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/0-cluster b/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/0-cluster deleted file mode 120000 index 3580937..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/0-cluster +++ /dev/null @@ -1 +0,0 @@ -../stress-split/0-cluster/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/1-kraken-install b/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/1-kraken-install deleted file mode 120000 index d4bcb5a..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/1-kraken-install +++ /dev/null @@ -1 +0,0 @@ -../stress-split/1-kraken-install/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/2-partial-upgrade b/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/2-partial-upgrade deleted file mode 120000 index ab35fc1..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/2-partial-upgrade +++ /dev/null @@ -1 +0,0 @@ -../stress-split/2-partial-upgrade/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/3-thrash/default.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/3-thrash/default.yaml deleted file mode 100644 index edae7b3..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/3-thrash/default.yaml +++ /dev/null @@ -1,25 +0,0 @@ -meta: -- desc: | - randomly kill and revive osd - small chance to increase the number of pgs -overrides: - ceph: - log-whitelist: - - but it is still running - - wrongly marked me down - - objects unfound and apparently lost - - log bound mismatch -tasks: -- parallel: - - stress-tasks -stress-tasks: -- thrashosds: - timeout: 1200 - chance_pgnum_grow: 1 - chance_pgpnum_fix: 1 - min_in: 4 - chance_thrash_cluster_full: 0 - chance_thrash_pg_upmap: 0 - chance_thrash_pg_upmap_items: 0 - chance_force_recovery: 0 -- print: "**** done thrashosds 3-thrash" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/4-ec-workload.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/4-ec-workload.yaml deleted file mode 100644 index c89551e..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/4-ec-workload.yaml +++ /dev/null @@ -1,22 +0,0 @@ -meta: -- desc: | - randomized correctness test for rados operations on an erasure coded pool -stress-tasks: - - rados: - clients: [client.0] - ops: 4000 - objects: 50 - ec_pool: true - write_append_excl: false - op_weights: - read: 100 - write: 0 - append: 100 - delete: 50 - snap_create: 50 - snap_remove: 50 - rollback: 50 - copy_from: 50 - setattr: 25 - rmattr: 25 - - print: "**** done rados ec task" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/5-finish-upgrade.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/5-finish-upgrade.yaml deleted file mode 120000 index a66a7dc..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/5-finish-upgrade.yaml +++ /dev/null @@ -1 +0,0 @@ -../stress-split/5-finish-upgrade.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/6-luminous-with-mgr.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/6-luminous-with-mgr.yaml deleted file mode 120000 index 01d44cc..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/6-luminous-with-mgr.yaml +++ /dev/null @@ -1 +0,0 @@ -../stress-split/6-luminous-with-mgr.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/7-final-workload.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/7-final-workload.yaml deleted file mode 100644 index 50a1465..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/7-final-workload.yaml +++ /dev/null @@ -1,35 +0,0 @@ -# -# k=3 implies a stripe_width of 1376*3 = 4128 which is different from -# the default value of 4096 It is also not a multiple of 1024*1024 and -# creates situations where rounding rules during recovery becomes -# necessary. -# -meta: -- desc: | - randomized correctness test for rados operations on an erasure coded pool - using the jerasure plugin with k=3 and m=1 -tasks: -- rados: - clients: [client.0] - ops: 4000 - objects: 50 - ec_pool: true - write_append_excl: false - erasure_code_profile: - name: jerasure31profile - plugin: jerasure - k: 3 - m: 1 - technique: reed_sol_van - crush-failure-domain: osd - op_weights: - read: 100 - write: 0 - append: 100 - delete: 50 - snap_create: 50 - snap_remove: 50 - rollback: 50 - copy_from: 50 - setattr: 25 - rmattr: 25 diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/distros b/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/distros deleted file mode 120000 index ca99fee..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/distros +++ /dev/null @@ -1 +0,0 @@ -../../../../distros/supported/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/objectstore b/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/objectstore deleted file mode 120000 index 016cbf9..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/objectstore +++ /dev/null @@ -1 +0,0 @@ -../stress-split/objectstore/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/thrashosds-health.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/thrashosds-health.yaml deleted file mode 120000 index e0426db..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split-erasure-code/thrashosds-health.yaml +++ /dev/null @@ -1 +0,0 @@ -../../../../tasks/thrashosds-health.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/% b/src/ceph/qa/suites/upgrade/kraken-x/stress-split/% deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/% +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/0-cluster/+ b/src/ceph/qa/suites/upgrade/kraken-x/stress-split/0-cluster/+ deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/0-cluster/+ +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/0-cluster/openstack.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split/0-cluster/openstack.yaml deleted file mode 100644 index a0d5c20..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/0-cluster/openstack.yaml +++ /dev/null @@ -1,6 +0,0 @@ -openstack: - - machine: - disk: 100 # GB - - volumes: # attached to each instance - count: 3 - size: 30 # GB diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/0-cluster/start.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split/0-cluster/start.yaml deleted file mode 100644 index b8a28f9..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/0-cluster/start.yaml +++ /dev/null @@ -1,27 +0,0 @@ -meta: -- desc: | - Run ceph on two nodes, - with a separate client-only node. - Use xfs beneath the osds. -overrides: - ceph: - fs: xfs - log-whitelist: - - overall HEALTH_ - - \(MON_DOWN\) - - \(MGR_DOWN\) - conf: - global: - enable experimental unrecoverable data corrupting features: "*" -roles: -- - mon.a - - mon.b - - mon.c - - mgr.x - - osd.0 - - osd.1 - - osd.2 -- - osd.3 - - osd.4 - - osd.5 -- - client.0 diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/1-kraken-install/kraken.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split/1-kraken-install/kraken.yaml deleted file mode 100644 index 145c2c8..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/1-kraken-install/kraken.yaml +++ /dev/null @@ -1,8 +0,0 @@ -meta: -- desc: install ceph/kraken latest -tasks: -- install: - branch: kraken -- print: "**** done install kraken" -- ceph: -- print: "**** done ceph" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/2-partial-upgrade/firsthalf.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split/2-partial-upgrade/firsthalf.yaml deleted file mode 100644 index 87fa1d5..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/2-partial-upgrade/firsthalf.yaml +++ /dev/null @@ -1,12 +0,0 @@ -meta: -- desc: | - install upgrade ceph/-x on one node only - 1st half - restart : osd.0,1,2 -tasks: -- install.upgrade: - osd.0: -- print: "**** done install.upgrade osd.0" -- ceph.restart: - daemons: [mon.a,mon.b,mon.c,mgr.x,osd.0,osd.1,osd.2] -- print: "**** done ceph.restart 1st half" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/3-thrash/default.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split/3-thrash/default.yaml deleted file mode 100644 index b3fddef..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/3-thrash/default.yaml +++ /dev/null @@ -1,25 +0,0 @@ -meta: -- desc: | - randomly kill and revive osd - small chance to increase the number of pgs -overrides: - ceph: - log-whitelist: - - but it is still running - - wrongly marked me down - - objects unfound and apparently lost - - log bound mismatch -tasks: -- parallel: - - stress-tasks -stress-tasks: -- thrashosds: - timeout: 1200 - chance_pgnum_grow: 1 - chance_pgpnum_fix: 1 - chance_thrash_cluster_full: 0 - chance_thrash_pg_upmap: 0 - chance_thrash_pg_upmap_items: 0 - disable_objectstore_tool_tests: true - chance_force_recovery: 0 -- print: "**** done thrashosds 3-thrash" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/4-workload/+ b/src/ceph/qa/suites/upgrade/kraken-x/stress-split/4-workload/+ deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/4-workload/+ +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/4-workload/radosbench.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split/4-workload/radosbench.yaml deleted file mode 100644 index 626ae8e..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/4-workload/radosbench.yaml +++ /dev/null @@ -1,40 +0,0 @@ -meta: -- desc: | - run randomized correctness test for rados operations - generate write load with rados bench -stress-tasks: -- full_sequential: - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 -- print: "**** done radosbench 7-workload" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/4-workload/rbd-cls.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split/4-workload/rbd-cls.yaml deleted file mode 100644 index 7f4b06b..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/4-workload/rbd-cls.yaml +++ /dev/null @@ -1,10 +0,0 @@ -meta: -- desc: | - run basic cls tests for rbd -stress-tasks: -- workunit: - branch: kraken - clients: - client.0: - - cls/test_cls_rbd.sh -- print: "**** done cls/test_cls_rbd.sh 5-workload" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/4-workload/rbd-import-export.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split/4-workload/rbd-import-export.yaml deleted file mode 100644 index b8b6ad3..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/4-workload/rbd-import-export.yaml +++ /dev/null @@ -1,12 +0,0 @@ -meta: -- desc: | - run basic import/export cli tests for rbd -stress-tasks: -- workunit: - branch: kraken - clients: - client.0: - - rbd/import_export.sh - env: - RBD_CREATE_ARGS: --new-format -- print: "**** done rbd/import_export.sh 5-workload" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/4-workload/rbd_api.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split/4-workload/rbd_api.yaml deleted file mode 100644 index a5ae1e5..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/4-workload/rbd_api.yaml +++ /dev/null @@ -1,10 +0,0 @@ -meta: -- desc: | - librbd C and C++ api tests -stress-tasks: -- workunit: - branch: kraken - clients: - client.0: - - rbd/test_librbd.sh -- print: "**** done rbd/test_librbd.sh 7-workload" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/4-workload/readwrite.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split/4-workload/readwrite.yaml deleted file mode 100644 index 41e34d6..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/4-workload/readwrite.yaml +++ /dev/null @@ -1,16 +0,0 @@ -meta: -- desc: | - randomized correctness test for rados operations on a replicated pool, - using only reads, writes, and deletes -stress-tasks: -- full_sequential: - - rados: - clients: [client.0] - ops: 4000 - objects: 500 - write_append_excl: false - op_weights: - read: 45 - write: 45 - delete: 10 -- print: "**** done rados/readwrite 5-workload" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/4-workload/snaps-few-objects.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split/4-workload/snaps-few-objects.yaml deleted file mode 100644 index f56d0de..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/4-workload/snaps-few-objects.yaml +++ /dev/null @@ -1,18 +0,0 @@ -meta: -- desc: | - randomized correctness test for rados operations on a replicated pool with snapshot operations -stress-tasks: -- full_sequential: - - rados: - clients: [client.0] - ops: 4000 - objects: 50 - write_append_excl: false - op_weights: - read: 100 - write: 100 - delete: 50 - snap_create: 50 - snap_remove: 50 - rollback: 50 -- print: "**** done rados/snaps-few-objects 5-workload" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/5-finish-upgrade.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split/5-finish-upgrade.yaml deleted file mode 100644 index 1d528cd..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/5-finish-upgrade.yaml +++ /dev/null @@ -1,9 +0,0 @@ -tasks: -- install.upgrade: - osd.3: - client.0: -- ceph.restart: - daemons: [osd.3, osd.4, osd.5] - wait-for-healthy: false - wait-for-osds-up: true - diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/6-luminous-with-mgr.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split/6-luminous-with-mgr.yaml deleted file mode 120000 index 5c72153..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/6-luminous-with-mgr.yaml +++ /dev/null @@ -1 +0,0 @@ -../../../../releases/luminous-with-mgr.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/7-final-workload/+ b/src/ceph/qa/suites/upgrade/kraken-x/stress-split/7-final-workload/+ deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/7-final-workload/+ +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/7-final-workload/rbd-python.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split/7-final-workload/rbd-python.yaml deleted file mode 100644 index 24c2644..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/7-final-workload/rbd-python.yaml +++ /dev/null @@ -1,10 +0,0 @@ -meta: -- desc: | - librbd python api tests -tasks: -- workunit: - branch: kraken - clients: - client.0: - - rbd/test_librbd_python.sh -- print: "**** done rbd/test_librbd_python.sh 9-workload" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/7-final-workload/rgw-swift.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split/7-final-workload/rgw-swift.yaml deleted file mode 100644 index 76e5d6f..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/7-final-workload/rgw-swift.yaml +++ /dev/null @@ -1,11 +0,0 @@ -meta: -- desc: | - swift api tests for rgw -tasks: -- rgw: - client.0: -- print: "**** done rgw 9-workload" -- swift: - client.0: - rgw_server: client.0 -- print: "**** done swift 9-workload" diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/7-final-workload/snaps-many-objects.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split/7-final-workload/snaps-many-objects.yaml deleted file mode 100644 index 805bf97..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/7-final-workload/snaps-many-objects.yaml +++ /dev/null @@ -1,16 +0,0 @@ -meta: -- desc: | - randomized correctness test for rados operations on a replicated pool with snapshot operations -tasks: -- rados: - clients: [client.0] - ops: 4000 - objects: 500 - write_append_excl: false - op_weights: - read: 100 - write: 100 - delete: 50 - snap_create: 50 - snap_remove: 50 - rollback: 50 diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/distros b/src/ceph/qa/suites/upgrade/kraken-x/stress-split/distros deleted file mode 120000 index ca99fee..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/distros +++ /dev/null @@ -1 +0,0 @@ -../../../../distros/supported/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/objectstore/bluestore.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split/objectstore/bluestore.yaml deleted file mode 120000 index d644598..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/objectstore/bluestore.yaml +++ /dev/null @@ -1 +0,0 @@ -../../../../../objectstore/bluestore.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/objectstore/filestore-xfs.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split/objectstore/filestore-xfs.yaml deleted file mode 120000 index 03750e5..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/objectstore/filestore-xfs.yaml +++ /dev/null @@ -1 +0,0 @@ -../../../../../objectstore/filestore-xfs.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/thrashosds-health.yaml b/src/ceph/qa/suites/upgrade/kraken-x/stress-split/thrashosds-health.yaml deleted file mode 120000 index e0426db..0000000 --- a/src/ceph/qa/suites/upgrade/kraken-x/stress-split/thrashosds-health.yaml +++ /dev/null @@ -1 +0,0 @@ -../../../../tasks/thrashosds-health.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/% b/src/ceph/qa/suites/upgrade/luminous-x/parallel/% deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/% +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/0-cluster/+ b/src/ceph/qa/suites/upgrade/luminous-x/parallel/0-cluster/+ deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/0-cluster/+ +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/0-cluster/openstack.yaml b/src/ceph/qa/suites/upgrade/luminous-x/parallel/0-cluster/openstack.yaml deleted file mode 100644 index f4d1349..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/0-cluster/openstack.yaml +++ /dev/null @@ -1,4 +0,0 @@ -openstack: - - volumes: # attached to each instance - count: 3 - size: 30 # GB diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/0-cluster/start.yaml b/src/ceph/qa/suites/upgrade/luminous-x/parallel/0-cluster/start.yaml deleted file mode 100644 index 3684b1e..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/0-cluster/start.yaml +++ /dev/null @@ -1,40 +0,0 @@ -meta: -- desc: | - Run ceph on two nodes, - with a separate client 0,1,2 third node. - Use xfs beneath the osds. - CephFS tests running on client 2,3 -roles: -- - mon.a - - mgr.x - - mds.a - - osd.0 - - osd.1 -- - mon.b - - mon.c - - osd.2 - - osd.3 -- - client.0 - - client.1 - - client.2 - - client.3 -- - client.4 -overrides: - ceph: - log-whitelist: - - scrub mismatch - - ScrubResult - - wrongly marked - - (POOL_APP_NOT_ENABLED) - - overall HEALTH_ - conf: - global: - enable experimental unrecoverable data corrupting features: "*" - mon: - mon warn on osd down out interval zero: false - osd: - osd_class_load_list: "cephfs hello journal lock log numops rbd refcount - replica_log rgw sdk statelog timeindex user version" - osd_class_default_list: "cephfs hello journal lock log numops rbd refcount - replica_log rgw sdk statelog timeindex user version" - fs: xfs diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/1-ceph-install/luminous.yaml b/src/ceph/qa/suites/upgrade/luminous-x/parallel/1-ceph-install/luminous.yaml deleted file mode 100644 index 3d57f79..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/1-ceph-install/luminous.yaml +++ /dev/null @@ -1,43 +0,0 @@ -meta: -- desc: | - install ceph/luminous latest - run workload and upgrade-sequence in parallel - upgrade the client node -tasks: -- install: - branch: luminous -- print: "**** done installing luminous" -- ceph: - log-whitelist: - - overall HEALTH_ - - \(FS_ - - \(MDS_ - - \(OSD_ - - \(MON_DOWN\) - - \(CACHE_POOL_ - - \(POOL_ - - \(MGR_DOWN\) - - \(PG_ - - \(SMALLER_PGP_NUM\) - - Monitor daemon marked osd - - Behind on trimming - - Manager daemon - conf: - global: - mon warn on pool no app: false -- exec: - osd.0: - - ceph osd require-osd-release luminous - - ceph osd set-require-min-compat-client luminous -- print: "**** done ceph" -- install.upgrade: - mon.a: - mon.b: -- print: "**** done install.upgrade both hosts" -- parallel: - - workload - - upgrade-sequence -- print: "**** done parallel" -- install.upgrade: - client.0: -- print: "**** done install.upgrade on client.0" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/2-workload/+ b/src/ceph/qa/suites/upgrade/luminous-x/parallel/2-workload/+ deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/2-workload/+ +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/2-workload/blogbench.yaml b/src/ceph/qa/suites/upgrade/luminous-x/parallel/2-workload/blogbench.yaml deleted file mode 100644 index 021fcc6..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/2-workload/blogbench.yaml +++ /dev/null @@ -1,14 +0,0 @@ -meta: -- desc: | - run a cephfs stress test - mount ceph-fuse on client.2 before running workunit -workload: - full_sequential: - - sequential: - - ceph-fuse: - - print: "**** done ceph-fuse 2-workload" - - workunit: - clients: - client.2: - - suites/blogbench.sh - - print: "**** done suites/blogbench.sh 2-workload" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/2-workload/ec-rados-default.yaml b/src/ceph/qa/suites/upgrade/luminous-x/parallel/2-workload/ec-rados-default.yaml deleted file mode 100644 index 5c5a958..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/2-workload/ec-rados-default.yaml +++ /dev/null @@ -1,24 +0,0 @@ -meta: -- desc: | - run run randomized correctness test for rados operations - on an erasure-coded pool -workload: - full_sequential: - - rados: - clients: [client.0] - ops: 4000 - objects: 50 - ec_pool: true - write_append_excl: false - op_weights: - read: 100 - write: 0 - append: 100 - delete: 50 - snap_create: 50 - snap_remove: 50 - rollback: 50 - copy_from: 50 - setattr: 25 - rmattr: 25 - - print: "**** done rados ec task" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/2-workload/rados_api.yaml b/src/ceph/qa/suites/upgrade/luminous-x/parallel/2-workload/rados_api.yaml deleted file mode 100644 index e4cc9f9..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/2-workload/rados_api.yaml +++ /dev/null @@ -1,11 +0,0 @@ -meta: -- desc: | - object class functional tests -workload: - full_sequential: - - workunit: - branch: luminous - clients: - client.0: - - cls - - print: "**** done cls 2-workload" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/2-workload/rados_loadgenbig.yaml b/src/ceph/qa/suites/upgrade/luminous-x/parallel/2-workload/rados_loadgenbig.yaml deleted file mode 100644 index 874a8c5..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/2-workload/rados_loadgenbig.yaml +++ /dev/null @@ -1,11 +0,0 @@ -meta: -- desc: | - generate read/write load with rados objects ranging from 1MB to 25MB -workload: - full_sequential: - - workunit: - branch: luminous - clients: - client.0: - - rados/load-gen-big.sh - - print: "**** done rados/load-gen-big.sh 2-workload" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/2-workload/test_rbd_api.yaml b/src/ceph/qa/suites/upgrade/luminous-x/parallel/2-workload/test_rbd_api.yaml deleted file mode 100644 index 81563c9..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/2-workload/test_rbd_api.yaml +++ /dev/null @@ -1,11 +0,0 @@ -meta: -- desc: | - librbd C and C++ api tests -workload: - full_sequential: - - workunit: - branch: luminous - clients: - client.0: - - rbd/test_librbd.sh - - print: "**** done rbd/test_librbd.sh 2-workload" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/2-workload/test_rbd_python.yaml b/src/ceph/qa/suites/upgrade/luminous-x/parallel/2-workload/test_rbd_python.yaml deleted file mode 100644 index e17207d..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/2-workload/test_rbd_python.yaml +++ /dev/null @@ -1,11 +0,0 @@ -meta: -- desc: | - librbd python api tests -workload: - full_sequential: - - workunit: - branch: luminous - clients: - client.0: - - rbd/test_librbd_python.sh - - print: "**** done rbd/test_librbd_python.sh 2-workload" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/3-upgrade-sequence/upgrade-all.yaml b/src/ceph/qa/suites/upgrade/luminous-x/parallel/3-upgrade-sequence/upgrade-all.yaml deleted file mode 100644 index cff3a68..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/3-upgrade-sequence/upgrade-all.yaml +++ /dev/null @@ -1,16 +0,0 @@ -meta: -- desc: | - upgrade the ceph cluster -upgrade-sequence: - sequential: - - ceph.restart: - daemons: [mon.a, mon.b, mon.c, mgr.x] - - ceph.restart: - daemons: [osd.0, osd.1, osd.2, osd.3] - wait-for-healthy: false - wait-for-osds-up: true - - ceph.restart: - daemons: [mds.a] - wait-for-healthy: false - wait-for-osds-up: true - - print: "**** done ceph.restart all" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/3-upgrade-sequence/upgrade-mon-osd-mds.yaml b/src/ceph/qa/suites/upgrade/luminous-x/parallel/3-upgrade-sequence/upgrade-mon-osd-mds.yaml deleted file mode 100644 index f197de6..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/3-upgrade-sequence/upgrade-mon-osd-mds.yaml +++ /dev/null @@ -1,35 +0,0 @@ -meta: -- desc: | - upgrade the ceph cluster, - upgrate in two steps - step one ordering: mon.a, osd.0, osd.1, mds.a - step two ordering: mon.b, mon.c, osd.2, osd.3 - ceph expected to be healthy state after each step -upgrade-sequence: - sequential: - - ceph.restart: - daemons: [mon.a] - wait-for-healthy: true - - sleep: - duration: 60 - - ceph.restart: - daemons: [mon.b, mon.c, mgr.x] - wait-for-healthy: true - - sleep: - duration: 60 - - ceph.restart: - daemons: [osd.0, osd.1] - wait-for-healthy: true - - sleep: - duration: 60 - - ceph.restart: [mds.a] - - sleep: - duration: 60 - - sleep: - duration: 60 - - ceph.restart: - daemons: [osd.2, osd.3] - wait-for-healthy: false - wait-for-osds-up: true - - sleep: - duration: 60 diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/+ b/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/+ deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/+ +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/blogbench.yaml b/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/blogbench.yaml deleted file mode 100644 index d2629c0..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/blogbench.yaml +++ /dev/null @@ -1,13 +0,0 @@ -meta: -- desc: | - run a cephfs stress test - mount ceph-fuse on client.3 before running workunit -tasks: -- sequential: - - ceph-fuse: - - print: "**** done ceph-fuse 5-final-workload" - - workunit: - clients: - client.3: - - suites/blogbench.sh - - print: "**** done suites/blogbench.sh 5-final-workload" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/rados-snaps-few-objects.yaml b/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/rados-snaps-few-objects.yaml deleted file mode 100644 index d8b3dcb..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/rados-snaps-few-objects.yaml +++ /dev/null @@ -1,17 +0,0 @@ -meta: -- desc: | - randomized correctness test for rados operations on a replicated pool with snapshots -tasks: - - rados: - clients: [client.1] - ops: 4000 - objects: 50 - write_append_excl: false - op_weights: - read: 100 - write: 100 - delete: 50 - snap_create: 50 - snap_remove: 50 - rollback: 50 - - print: "**** done rados 4-final-workload" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/rados_loadgenmix.yaml b/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/rados_loadgenmix.yaml deleted file mode 100644 index 922a9da..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/rados_loadgenmix.yaml +++ /dev/null @@ -1,9 +0,0 @@ -meta: -- desc: | - generate read/write load with rados objects ranging from 1 byte to 1MB -tasks: - - workunit: - clients: - client.1: - - rados/load-gen-mix.sh - - print: "**** done rados/load-gen-mix.sh 4-final-workload" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/rados_mon_thrash.yaml b/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/rados_mon_thrash.yaml deleted file mode 100644 index a42b7d2..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/rados_mon_thrash.yaml +++ /dev/null @@ -1,18 +0,0 @@ -meta: -- desc: | - librados C and C++ api tests -overrides: - ceph: - log-whitelist: - - reached quota -tasks: - - mon_thrash: - revive_delay: 20 - thrash_delay: 1 - - print: "**** done mon_thrash 4-final-workload" - - workunit: - branch: luminous - clients: - client.1: - - rados/test.sh - - print: "**** done rados/test.sh 4-final-workload" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/rbd_cls.yaml b/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/rbd_cls.yaml deleted file mode 100644 index aaf0a37..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/rbd_cls.yaml +++ /dev/null @@ -1,9 +0,0 @@ -meta: -- desc: | - rbd object class functional tests -tasks: - - workunit: - clients: - client.1: - - cls/test_cls_rbd.sh - - print: "**** done cls/test_cls_rbd.sh 4-final-workload" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/rbd_import_export_no_upgrated.yaml b/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/rbd_import_export_no_upgrated.yaml deleted file mode 100644 index 5de8a23..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/rbd_import_export_no_upgrated.yaml +++ /dev/null @@ -1,13 +0,0 @@ -meta: -- desc: | - run basic import/export cli tests for rbd - on NO upgrated client -tasks: - - workunit: - branch: luminous - clients: - client.4: - - rbd/import_export.sh - env: - RBD_CREATE_ARGS: --new-format - - print: "**** done rbd/import_export.sh 4-final-workload on NO upgrated client" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/rbd_import_export_upgrated.yaml b/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/rbd_import_export_upgrated.yaml deleted file mode 100644 index 2c7c484..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/rbd_import_export_upgrated.yaml +++ /dev/null @@ -1,12 +0,0 @@ -meta: -- desc: | - run basic import/export cli tests for rbd - on upgrated client -tasks: - - workunit: - clients: - client.1: - - rbd/import_export.sh - env: - RBD_CREATE_ARGS: --new-format - - print: "**** done rbd/import_export.sh 4-final-workload on upgrated client" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/rgw_swift.yaml b/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/rgw_swift.yaml deleted file mode 100644 index 7a7659f..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/5-final-workload/rgw_swift.yaml +++ /dev/null @@ -1,13 +0,0 @@ -meta: -- desc: | - swift api tests for rgw -overrides: - rgw: - frontend: civetweb -tasks: - - rgw: [client.1] - - print: "**** done rgw 4-final-workload" - - swift: - client.1: - rgw_server: client.1 - - print: "**** done swift 4-final-workload" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/distros b/src/ceph/qa/suites/upgrade/luminous-x/parallel/distros deleted file mode 120000 index ca99fee..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/distros +++ /dev/null @@ -1 +0,0 @@ -../../../../distros/supported/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/luminous-x/parallel/objectstore b/src/ceph/qa/suites/upgrade/luminous-x/parallel/objectstore deleted file mode 120000 index 016cbf9..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/parallel/objectstore +++ /dev/null @@ -1 +0,0 @@ -../stress-split/objectstore/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/luminous-x/point-to-point-x/% b/src/ceph/qa/suites/upgrade/luminous-x/point-to-point-x/% deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/point-to-point-x/% +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/luminous-x/point-to-point-x/distros/centos_latest.yaml b/src/ceph/qa/suites/upgrade/luminous-x/point-to-point-x/distros/centos_latest.yaml deleted file mode 120000 index b5973b9..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/point-to-point-x/distros/centos_latest.yaml +++ /dev/null @@ -1 +0,0 @@ -../../../../../distros/supported/centos_latest.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/luminous-x/point-to-point-x/distros/ubuntu_latest.yaml b/src/ceph/qa/suites/upgrade/luminous-x/point-to-point-x/distros/ubuntu_latest.yaml deleted file mode 120000 index cc5b15b..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/point-to-point-x/distros/ubuntu_latest.yaml +++ /dev/null @@ -1 +0,0 @@ -../../../../../distros/supported/ubuntu_latest.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/luminous-x/point-to-point-x/point-to-point-upgrade.yaml b/src/ceph/qa/suites/upgrade/luminous-x/point-to-point-x/point-to-point-upgrade.yaml deleted file mode 100644 index 4c81c34..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/point-to-point-x/point-to-point-upgrade.yaml +++ /dev/null @@ -1,225 +0,0 @@ -meta: -- desc: | - Run ceph on two nodes, using one of them as a client, - with a separate client-only node. - Use xfs beneath the osds. - install ceph/luminous v12.2.2 point version - run workload and upgrade-sequence in parallel - install ceph/luminous latest version - run workload and upgrade-sequence in parallel - install ceph/-x version (luminous or master/mimic) - run workload and upgrade-sequence in parallel -overrides: - ceph: - log-whitelist: - - reached quota - - scrub - - osd_map_max_advance - - wrongly marked - fs: xfs - conf: - mon: - mon debug unsafe allow tier with nonempty snaps: true - mon warn on pool no app: false - osd: - osd map max advance: 1000 - osd_class_load_list: "cephfs hello journal lock log numops rbd refcount - replica_log rgw sdk statelog timeindex user version" - osd_class_default_list: "cephfs hello journal lock log numops rbd refcount - replica_log rgw sdk statelog timeindex user version" - client: - rgw_crypt_require_ssl: false - rgw crypt s3 kms encryption keys: testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo= testkey-2=aWIKTWFrZWZpbGUKbWFuCm91dApzcmMKVGVzdGluZwo= -roles: -- - mon.a - - mds.a - - osd.0 - - osd.1 - - osd.2 - - mgr.x -- - mon.b - - mon.c - - osd.3 - - osd.4 - - osd.5 - - client.0 -- - client.1 -openstack: -- volumes: # attached to each instance - count: 3 - size: 30 # GB -tasks: -- print: "**** v12.2.2 about to install" -- install: - tag: v12.2.2 - # line below can be removed its from jewel test - #exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev', 'librgw2'] -- print: "**** done v12.2.2 install" -- ceph: - fs: xfs - add_osds_to_crush: true -- print: "**** done ceph xfs" -- sequential: - - workload -- print: "**** done workload" -- install.upgrade: - #exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] - mon.a: - branch: luminous - mon.b: - branch: luminous - # Note that client.a IS NOT upgraded at this point -- parallel: - - workload_luminous - - upgrade-sequence_luminous -- print: "**** done parallel luminous branch" -- install.upgrade: - #exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] - client.1: - branch: luminous -- print: "**** done branch: luminous install.upgrade on client.1" -- install.upgrade: - mon.a: - mon.b: -- print: "**** done branch: -x install.upgrade on mon.a and mon.b" -- parallel: - - workload_x - - upgrade-sequence_x -- print: "**** done parallel -x branch" -- exec: - osd.0: - - ceph osd set-require-min-compat-client luminous -# Run librados tests on the -x upgraded cluster -- install.upgrade: - client.1: -- workunit: - branch: luminous - clients: - client.1: - - rados/test.sh - - cls -- print: "**** done final test on -x cluster" -####################### -workload: - sequential: - - workunit: - clients: - client.0: - - suites/blogbench.sh -workload_luminous: - full_sequential: - - workunit: - branch: luminous - clients: - client.1: - - rados/test.sh - - cls - - print: "**** done rados/test.sh & cls workload_luminous" - - sequential: - - rgw: [client.0] - - print: "**** done rgw workload_luminous" - - s3tests: - client.0: - force-branch: ceph-luminous - rgw_server: client.0 - scan_for_encryption_keys: false - - print: "**** done s3tests workload_luminous" -upgrade-sequence_luminous: - sequential: - - print: "**** done branch: luminous install.upgrade" - - ceph.restart: [mds.a] - - sleep: - duration: 60 - - ceph.restart: [osd.0] - - sleep: - duration: 30 - - ceph.restart: [osd.1] - - sleep: - duration: 30 - - ceph.restart: [osd.2] - - sleep: - duration: 30 - - ceph.restart: [osd.3] - - sleep: - duration: 30 - - ceph.restart: [osd.4] - - sleep: - duration: 30 - - ceph.restart: [osd.5] - - sleep: - duration: 60 - - ceph.restart: [mon.a] - - sleep: - duration: 60 - - ceph.restart: [mon.b] - - sleep: - duration: 60 - - ceph.restart: [mon.c] - - sleep: - duration: 60 - - print: "**** done ceph.restart all luminous branch mds/osd/mon" -workload_x: - sequential: - - workunit: - branch: luminous - clients: - client.1: - - rados/test.sh - - cls - - print: "**** done rados/test.sh & cls workload_x NOT upgraded client" - - workunit: - branch: luminous - clients: - client.0: - - rados/test.sh - - cls - - print: "**** done rados/test.sh & cls workload_x upgraded client" - - rgw: [client.1] - - print: "**** done rgw workload_x" - - s3tests: - client.1: - force-branch: ceph-luminous - rgw_server: client.1 - scan_for_encryption_keys: false - - print: "**** done s3tests workload_x" -upgrade-sequence_x: - sequential: - - ceph.restart: [mds.a] - - sleep: - duration: 60 - - ceph.restart: [mon.a] - - sleep: - duration: 60 - - ceph.restart: [mon.b] - - sleep: - duration: 60 - - ceph.restart: [mon.c] - - sleep: - duration: 60 - - ceph.restart: [osd.0] - - sleep: - duration: 30 - - ceph.restart: [osd.1] - - sleep: - duration: 30 - - ceph.restart: [osd.2] - - sleep: - duration: 30 - - ceph.restart: [osd.3] - - sleep: - duration: 30 - - ceph.restart: [osd.4] - - sleep: - duration: 30 - - ceph.restart: - daemons: [osd.5] - wait-for-healthy: false - wait-for-up-osds: true - - ceph.restart: - daemons: [mgr.x] - wait-for-healthy: false - - exec: - osd.0: - - ceph osd require-osd-release luminous - - ceph.healthy: - - print: "**** done ceph.restart all -x branch mds/osd/mon" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/% b/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/% deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/% +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/0-cluster b/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/0-cluster deleted file mode 120000 index 3580937..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/0-cluster +++ /dev/null @@ -1 +0,0 @@ -../stress-split/0-cluster/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/1-ceph-install b/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/1-ceph-install deleted file mode 120000 index 0479ac5..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/1-ceph-install +++ /dev/null @@ -1 +0,0 @@ -../stress-split/1-ceph-install/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/2-partial-upgrade b/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/2-partial-upgrade deleted file mode 120000 index ab35fc1..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/2-partial-upgrade +++ /dev/null @@ -1 +0,0 @@ -../stress-split/2-partial-upgrade/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/3-thrash/default.yaml b/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/3-thrash/default.yaml deleted file mode 100644 index edae7b3..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/3-thrash/default.yaml +++ /dev/null @@ -1,25 +0,0 @@ -meta: -- desc: | - randomly kill and revive osd - small chance to increase the number of pgs -overrides: - ceph: - log-whitelist: - - but it is still running - - wrongly marked me down - - objects unfound and apparently lost - - log bound mismatch -tasks: -- parallel: - - stress-tasks -stress-tasks: -- thrashosds: - timeout: 1200 - chance_pgnum_grow: 1 - chance_pgpnum_fix: 1 - min_in: 4 - chance_thrash_cluster_full: 0 - chance_thrash_pg_upmap: 0 - chance_thrash_pg_upmap_items: 0 - chance_force_recovery: 0 -- print: "**** done thrashosds 3-thrash" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/4-ec-workload.yaml b/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/4-ec-workload.yaml deleted file mode 100644 index c89551e..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/4-ec-workload.yaml +++ /dev/null @@ -1,22 +0,0 @@ -meta: -- desc: | - randomized correctness test for rados operations on an erasure coded pool -stress-tasks: - - rados: - clients: [client.0] - ops: 4000 - objects: 50 - ec_pool: true - write_append_excl: false - op_weights: - read: 100 - write: 0 - append: 100 - delete: 50 - snap_create: 50 - snap_remove: 50 - rollback: 50 - copy_from: 50 - setattr: 25 - rmattr: 25 - - print: "**** done rados ec task" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/5-finish-upgrade.yaml b/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/5-finish-upgrade.yaml deleted file mode 120000 index a66a7dc..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/5-finish-upgrade.yaml +++ /dev/null @@ -1 +0,0 @@ -../stress-split/5-finish-upgrade.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/7-final-workload.yaml b/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/7-final-workload.yaml deleted file mode 100644 index 50a1465..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/7-final-workload.yaml +++ /dev/null @@ -1,35 +0,0 @@ -# -# k=3 implies a stripe_width of 1376*3 = 4128 which is different from -# the default value of 4096 It is also not a multiple of 1024*1024 and -# creates situations where rounding rules during recovery becomes -# necessary. -# -meta: -- desc: | - randomized correctness test for rados operations on an erasure coded pool - using the jerasure plugin with k=3 and m=1 -tasks: -- rados: - clients: [client.0] - ops: 4000 - objects: 50 - ec_pool: true - write_append_excl: false - erasure_code_profile: - name: jerasure31profile - plugin: jerasure - k: 3 - m: 1 - technique: reed_sol_van - crush-failure-domain: osd - op_weights: - read: 100 - write: 0 - append: 100 - delete: 50 - snap_create: 50 - snap_remove: 50 - rollback: 50 - copy_from: 50 - setattr: 25 - rmattr: 25 diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/distros b/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/distros deleted file mode 120000 index ca99fee..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/distros +++ /dev/null @@ -1 +0,0 @@ -../../../../distros/supported/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/objectstore b/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/objectstore deleted file mode 120000 index 016cbf9..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/objectstore +++ /dev/null @@ -1 +0,0 @@ -../stress-split/objectstore/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/thrashosds-health.yaml b/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/thrashosds-health.yaml deleted file mode 120000 index e0426db..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split-erasure-code/thrashosds-health.yaml +++ /dev/null @@ -1 +0,0 @@ -../../../../tasks/thrashosds-health.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/% b/src/ceph/qa/suites/upgrade/luminous-x/stress-split/% deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/% +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/0-cluster/+ b/src/ceph/qa/suites/upgrade/luminous-x/stress-split/0-cluster/+ deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/0-cluster/+ +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/0-cluster/openstack.yaml b/src/ceph/qa/suites/upgrade/luminous-x/stress-split/0-cluster/openstack.yaml deleted file mode 100644 index a0d5c20..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/0-cluster/openstack.yaml +++ /dev/null @@ -1,6 +0,0 @@ -openstack: - - machine: - disk: 100 # GB - - volumes: # attached to each instance - count: 3 - size: 30 # GB diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/0-cluster/start.yaml b/src/ceph/qa/suites/upgrade/luminous-x/stress-split/0-cluster/start.yaml deleted file mode 100644 index e3ad918..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/0-cluster/start.yaml +++ /dev/null @@ -1,29 +0,0 @@ -meta: -- desc: | - Run ceph on two nodes, - with a separate client-only node. - Use xfs beneath the osds. -overrides: - ceph: - fs: xfs - log-whitelist: - - overall HEALTH_ - - \(MON_DOWN\) - - \(MGR_DOWN\) - conf: - global: - enable experimental unrecoverable data corrupting features: "*" - mon: - mon warn on osd down out interval zero: false -roles: -- - mon.a - - mon.b - - mon.c - - mgr.x - - osd.0 - - osd.1 - - osd.2 -- - osd.3 - - osd.4 - - osd.5 -- - client.0 diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/1-ceph-install/luminous.yaml b/src/ceph/qa/suites/upgrade/luminous-x/stress-split/1-ceph-install/luminous.yaml deleted file mode 100644 index 2230525..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/1-ceph-install/luminous.yaml +++ /dev/null @@ -1,17 +0,0 @@ -meta: -- desc: install ceph/luminous latest -tasks: -- install: - branch: luminous -- print: "**** done install luminous" -- ceph: -- exec: - osd.0: - - ceph osd require-osd-release luminous - - ceph osd set-require-min-compat-client luminous -- print: "**** done ceph " -overrides: - ceph: - conf: - mon: - mon warn on osd down out interval zero: false diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/2-partial-upgrade/firsthalf.yaml b/src/ceph/qa/suites/upgrade/luminous-x/stress-split/2-partial-upgrade/firsthalf.yaml deleted file mode 100644 index 87fa1d5..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/2-partial-upgrade/firsthalf.yaml +++ /dev/null @@ -1,12 +0,0 @@ -meta: -- desc: | - install upgrade ceph/-x on one node only - 1st half - restart : osd.0,1,2 -tasks: -- install.upgrade: - osd.0: -- print: "**** done install.upgrade osd.0" -- ceph.restart: - daemons: [mon.a,mon.b,mon.c,mgr.x,osd.0,osd.1,osd.2] -- print: "**** done ceph.restart 1st half" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/3-thrash/default.yaml b/src/ceph/qa/suites/upgrade/luminous-x/stress-split/3-thrash/default.yaml deleted file mode 100644 index b3fddef..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/3-thrash/default.yaml +++ /dev/null @@ -1,25 +0,0 @@ -meta: -- desc: | - randomly kill and revive osd - small chance to increase the number of pgs -overrides: - ceph: - log-whitelist: - - but it is still running - - wrongly marked me down - - objects unfound and apparently lost - - log bound mismatch -tasks: -- parallel: - - stress-tasks -stress-tasks: -- thrashosds: - timeout: 1200 - chance_pgnum_grow: 1 - chance_pgpnum_fix: 1 - chance_thrash_cluster_full: 0 - chance_thrash_pg_upmap: 0 - chance_thrash_pg_upmap_items: 0 - disable_objectstore_tool_tests: true - chance_force_recovery: 0 -- print: "**** done thrashosds 3-thrash" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/4-workload/+ b/src/ceph/qa/suites/upgrade/luminous-x/stress-split/4-workload/+ deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/4-workload/+ +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/4-workload/radosbench.yaml b/src/ceph/qa/suites/upgrade/luminous-x/stress-split/4-workload/radosbench.yaml deleted file mode 100644 index 626ae8e..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/4-workload/radosbench.yaml +++ /dev/null @@ -1,40 +0,0 @@ -meta: -- desc: | - run randomized correctness test for rados operations - generate write load with rados bench -stress-tasks: -- full_sequential: - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 - - radosbench: - clients: [client.0] - time: 150 -- print: "**** done radosbench 7-workload" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/4-workload/rbd-cls.yaml b/src/ceph/qa/suites/upgrade/luminous-x/stress-split/4-workload/rbd-cls.yaml deleted file mode 100644 index f8cc4d8..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/4-workload/rbd-cls.yaml +++ /dev/null @@ -1,10 +0,0 @@ -meta: -- desc: | - run basic cls tests for rbd -stress-tasks: -- workunit: - branch: luminous - clients: - client.0: - - cls/test_cls_rbd.sh -- print: "**** done cls/test_cls_rbd.sh 5-workload" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/4-workload/rbd-import-export.yaml b/src/ceph/qa/suites/upgrade/luminous-x/stress-split/4-workload/rbd-import-export.yaml deleted file mode 100644 index 30a677a..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/4-workload/rbd-import-export.yaml +++ /dev/null @@ -1,12 +0,0 @@ -meta: -- desc: | - run basic import/export cli tests for rbd -stress-tasks: -- workunit: - branch: luminous - clients: - client.0: - - rbd/import_export.sh - env: - RBD_CREATE_ARGS: --new-format -- print: "**** done rbd/import_export.sh 5-workload" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/4-workload/rbd_api.yaml b/src/ceph/qa/suites/upgrade/luminous-x/stress-split/4-workload/rbd_api.yaml deleted file mode 100644 index 9079aa3..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/4-workload/rbd_api.yaml +++ /dev/null @@ -1,10 +0,0 @@ -meta: -- desc: | - librbd C and C++ api tests -stress-tasks: -- workunit: - branch: luminous - clients: - client.0: - - rbd/test_librbd.sh -- print: "**** done rbd/test_librbd.sh 7-workload" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/4-workload/readwrite.yaml b/src/ceph/qa/suites/upgrade/luminous-x/stress-split/4-workload/readwrite.yaml deleted file mode 100644 index 41e34d6..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/4-workload/readwrite.yaml +++ /dev/null @@ -1,16 +0,0 @@ -meta: -- desc: | - randomized correctness test for rados operations on a replicated pool, - using only reads, writes, and deletes -stress-tasks: -- full_sequential: - - rados: - clients: [client.0] - ops: 4000 - objects: 500 - write_append_excl: false - op_weights: - read: 45 - write: 45 - delete: 10 -- print: "**** done rados/readwrite 5-workload" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/4-workload/snaps-few-objects.yaml b/src/ceph/qa/suites/upgrade/luminous-x/stress-split/4-workload/snaps-few-objects.yaml deleted file mode 100644 index f56d0de..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/4-workload/snaps-few-objects.yaml +++ /dev/null @@ -1,18 +0,0 @@ -meta: -- desc: | - randomized correctness test for rados operations on a replicated pool with snapshot operations -stress-tasks: -- full_sequential: - - rados: - clients: [client.0] - ops: 4000 - objects: 50 - write_append_excl: false - op_weights: - read: 100 - write: 100 - delete: 50 - snap_create: 50 - snap_remove: 50 - rollback: 50 -- print: "**** done rados/snaps-few-objects 5-workload" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/5-finish-upgrade.yaml b/src/ceph/qa/suites/upgrade/luminous-x/stress-split/5-finish-upgrade.yaml deleted file mode 100644 index 1d528cd..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/5-finish-upgrade.yaml +++ /dev/null @@ -1,9 +0,0 @@ -tasks: -- install.upgrade: - osd.3: - client.0: -- ceph.restart: - daemons: [osd.3, osd.4, osd.5] - wait-for-healthy: false - wait-for-osds-up: true - diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/7-final-workload/+ b/src/ceph/qa/suites/upgrade/luminous-x/stress-split/7-final-workload/+ deleted file mode 100644 index e69de29..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/7-final-workload/+ +++ /dev/null diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/7-final-workload/rbd-python.yaml b/src/ceph/qa/suites/upgrade/luminous-x/stress-split/7-final-workload/rbd-python.yaml deleted file mode 100644 index 92fe658..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/7-final-workload/rbd-python.yaml +++ /dev/null @@ -1,10 +0,0 @@ -meta: -- desc: | - librbd python api tests -tasks: -- workunit: - branch: luminous - clients: - client.0: - - rbd/test_librbd_python.sh -- print: "**** done rbd/test_librbd_python.sh 9-workload" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/7-final-workload/rgw-swift.yaml b/src/ceph/qa/suites/upgrade/luminous-x/stress-split/7-final-workload/rgw-swift.yaml deleted file mode 100644 index 76e5d6f..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/7-final-workload/rgw-swift.yaml +++ /dev/null @@ -1,11 +0,0 @@ -meta: -- desc: | - swift api tests for rgw -tasks: -- rgw: - client.0: -- print: "**** done rgw 9-workload" -- swift: - client.0: - rgw_server: client.0 -- print: "**** done swift 9-workload" diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/7-final-workload/snaps-many-objects.yaml b/src/ceph/qa/suites/upgrade/luminous-x/stress-split/7-final-workload/snaps-many-objects.yaml deleted file mode 100644 index 805bf97..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/7-final-workload/snaps-many-objects.yaml +++ /dev/null @@ -1,16 +0,0 @@ -meta: -- desc: | - randomized correctness test for rados operations on a replicated pool with snapshot operations -tasks: -- rados: - clients: [client.0] - ops: 4000 - objects: 500 - write_append_excl: false - op_weights: - read: 100 - write: 100 - delete: 50 - snap_create: 50 - snap_remove: 50 - rollback: 50 diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/distros b/src/ceph/qa/suites/upgrade/luminous-x/stress-split/distros deleted file mode 120000 index ca99fee..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/distros +++ /dev/null @@ -1 +0,0 @@ -../../../../distros/supported/
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/objectstore/bluestore.yaml b/src/ceph/qa/suites/upgrade/luminous-x/stress-split/objectstore/bluestore.yaml deleted file mode 120000 index d644598..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/objectstore/bluestore.yaml +++ /dev/null @@ -1 +0,0 @@ -../../../../../objectstore/bluestore.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/objectstore/filestore-xfs.yaml b/src/ceph/qa/suites/upgrade/luminous-x/stress-split/objectstore/filestore-xfs.yaml deleted file mode 120000 index 03750e5..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/objectstore/filestore-xfs.yaml +++ /dev/null @@ -1 +0,0 @@ -../../../../../objectstore/filestore-xfs.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/thrashosds-health.yaml b/src/ceph/qa/suites/upgrade/luminous-x/stress-split/thrashosds-health.yaml deleted file mode 120000 index e0426db..0000000 --- a/src/ceph/qa/suites/upgrade/luminous-x/stress-split/thrashosds-health.yaml +++ /dev/null @@ -1 +0,0 @@ -../../../../tasks/thrashosds-health.yaml
\ No newline at end of file |