diff options
Diffstat (limited to 'src/ceph/qa/suites/rados/thrash')
43 files changed, 506 insertions, 0 deletions
diff --git a/src/ceph/qa/suites/rados/thrash/% b/src/ceph/qa/suites/rados/thrash/% new file mode 100644 index 0000000..e69de29 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/% diff --git a/src/ceph/qa/suites/rados/thrash/0-size-min-size-overrides/2-size-1-min-size.yaml b/src/ceph/qa/suites/rados/thrash/0-size-min-size-overrides/2-size-1-min-size.yaml new file mode 120000 index 0000000..4c817a6 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/0-size-min-size-overrides/2-size-1-min-size.yaml @@ -0,0 +1 @@ +../../../../overrides/2-size-1-min-size.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/rados/thrash/0-size-min-size-overrides/2-size-2-min-size.yaml b/src/ceph/qa/suites/rados/thrash/0-size-min-size-overrides/2-size-2-min-size.yaml new file mode 120000 index 0000000..c429b07 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/0-size-min-size-overrides/2-size-2-min-size.yaml @@ -0,0 +1 @@ +../../../../overrides/2-size-2-min-size.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/rados/thrash/0-size-min-size-overrides/3-size-2-min-size.yaml b/src/ceph/qa/suites/rados/thrash/0-size-min-size-overrides/3-size-2-min-size.yaml new file mode 120000 index 0000000..8d529f0 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/0-size-min-size-overrides/3-size-2-min-size.yaml @@ -0,0 +1 @@ +../../../../overrides/3-size-2-min-size.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/rados/thrash/1-pg-log-overrides/normal_pg_log.yaml b/src/ceph/qa/suites/rados/thrash/1-pg-log-overrides/normal_pg_log.yaml new file mode 100644 index 0000000..e69de29 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/1-pg-log-overrides/normal_pg_log.yaml diff --git a/src/ceph/qa/suites/rados/thrash/1-pg-log-overrides/short_pg_log.yaml b/src/ceph/qa/suites/rados/thrash/1-pg-log-overrides/short_pg_log.yaml new file mode 120000 index 0000000..62010f4 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/1-pg-log-overrides/short_pg_log.yaml @@ -0,0 +1 @@ +../../../../overrides/short_pg_log.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/rados/thrash/backoff/normal.yaml b/src/ceph/qa/suites/rados/thrash/backoff/normal.yaml new file mode 100644 index 0000000..e69de29 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/backoff/normal.yaml diff --git a/src/ceph/qa/suites/rados/thrash/backoff/peering.yaml b/src/ceph/qa/suites/rados/thrash/backoff/peering.yaml new file mode 100644 index 0000000..66d0611 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/backoff/peering.yaml @@ -0,0 +1,5 @@ +overrides: + ceph: + conf: + osd: + osd backoff on peering: true diff --git a/src/ceph/qa/suites/rados/thrash/backoff/peering_and_degraded.yaml b/src/ceph/qa/suites/rados/thrash/backoff/peering_and_degraded.yaml new file mode 100644 index 0000000..e610990 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/backoff/peering_and_degraded.yaml @@ -0,0 +1,6 @@ +overrides: + ceph: + conf: + osd: + osd backoff on peering: true + osd backoff on degraded: true diff --git a/src/ceph/qa/suites/rados/thrash/ceph.yaml b/src/ceph/qa/suites/rados/thrash/ceph.yaml new file mode 100644 index 0000000..2030acb --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/ceph.yaml @@ -0,0 +1,3 @@ +tasks: +- install: +- ceph: diff --git a/src/ceph/qa/suites/rados/thrash/clusters/+ b/src/ceph/qa/suites/rados/thrash/clusters/+ new file mode 100644 index 0000000..e69de29 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/clusters/+ diff --git a/src/ceph/qa/suites/rados/thrash/clusters/fixed-2.yaml b/src/ceph/qa/suites/rados/thrash/clusters/fixed-2.yaml new file mode 120000 index 0000000..cd0791a --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/clusters/fixed-2.yaml @@ -0,0 +1 @@ +../../../../clusters/fixed-2.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/rados/thrash/clusters/openstack.yaml b/src/ceph/qa/suites/rados/thrash/clusters/openstack.yaml new file mode 100644 index 0000000..b0f3b9b --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/clusters/openstack.yaml @@ -0,0 +1,4 @@ +openstack: + - volumes: # attached to each instance + count: 4 + size: 30 # GB diff --git a/src/ceph/qa/suites/rados/thrash/d-require-luminous/at-end.yaml b/src/ceph/qa/suites/rados/thrash/d-require-luminous/at-end.yaml new file mode 100644 index 0000000..ef998cc --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/d-require-luminous/at-end.yaml @@ -0,0 +1,33 @@ +# do not require luminous osds at mkfs time; only set flag at +# the end of the test run, then do a final scrub (to convert any +# legacy snapsets), and verify we are healthy. +tasks: +- full_sequential_finally: + - exec: + mon.a: + - ceph osd require-osd-release luminous + - ceph osd pool application enable base rados || true +# make sure osds have latest map + - rados -p rbd bench 5 write -b 4096 + - ceph.healthy: + - ceph.osd_scrub_pgs: + cluster: ceph + - exec: + mon.a: + - sleep 15 + - ceph osd dump | grep purged_snapdirs + - ceph pg dump -f json-pretty + - "ceph pg dump sum -f json-pretty | grep num_legacy_snapsets | head -1 | grep ': 0'" +overrides: + ceph: + conf: + global: + mon debug no require luminous: true + +# setting luminous triggers peering, which *might* trigger health alerts + log-whitelist: + - overall HEALTH_ + - \(PG_AVAILABILITY\) + - \(PG_DEGRADED\) + thrashosds: + chance_thrash_cluster_full: 0 diff --git a/src/ceph/qa/suites/rados/thrash/d-require-luminous/at-mkfs-balancer-crush-compat.yaml b/src/ceph/qa/suites/rados/thrash/d-require-luminous/at-mkfs-balancer-crush-compat.yaml new file mode 100644 index 0000000..9eb7143 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/d-require-luminous/at-mkfs-balancer-crush-compat.yaml @@ -0,0 +1,11 @@ +overrides: + ceph: + conf: + mgr: + debug osd: 20 +tasks: +- exec: + mon.a: + - while ! ceph balancer status ; do sleep 1 ; done + - ceph balancer mode crush-compat + - ceph balancer on diff --git a/src/ceph/qa/suites/rados/thrash/d-require-luminous/at-mkfs-balancer-upmap.yaml b/src/ceph/qa/suites/rados/thrash/d-require-luminous/at-mkfs-balancer-upmap.yaml new file mode 100644 index 0000000..a1e0afe --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/d-require-luminous/at-mkfs-balancer-upmap.yaml @@ -0,0 +1,11 @@ +overrides: + ceph: + conf: + mgr: + debug osd: 20 +tasks: +- exec: + mon.a: + - while ! ceph balancer status ; do sleep 1 ; done + - ceph balancer mode upmap + - ceph balancer on diff --git a/src/ceph/qa/suites/rados/thrash/d-require-luminous/at-mkfs.yaml b/src/ceph/qa/suites/rados/thrash/d-require-luminous/at-mkfs.yaml new file mode 100644 index 0000000..e69de29 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/d-require-luminous/at-mkfs.yaml diff --git a/src/ceph/qa/suites/rados/thrash/msgr b/src/ceph/qa/suites/rados/thrash/msgr new file mode 120000 index 0000000..b29ecdd --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/msgr @@ -0,0 +1 @@ +../basic/msgr
\ No newline at end of file diff --git a/src/ceph/qa/suites/rados/thrash/msgr-failures/fastclose.yaml b/src/ceph/qa/suites/rados/thrash/msgr-failures/fastclose.yaml new file mode 100644 index 0000000..77fd730 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/msgr-failures/fastclose.yaml @@ -0,0 +1,6 @@ +overrides: + ceph: + conf: + global: + ms inject socket failures: 2500 + ms tcp read timeout: 5 diff --git a/src/ceph/qa/suites/rados/thrash/msgr-failures/few.yaml b/src/ceph/qa/suites/rados/thrash/msgr-failures/few.yaml new file mode 100644 index 0000000..477bffe --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/msgr-failures/few.yaml @@ -0,0 +1,7 @@ +overrides: + ceph: + conf: + global: + ms inject socket failures: 5000 + osd: + osd heartbeat use min delay socket: true diff --git a/src/ceph/qa/suites/rados/thrash/msgr-failures/osd-delay.yaml b/src/ceph/qa/suites/rados/thrash/msgr-failures/osd-delay.yaml new file mode 100644 index 0000000..a33ba89 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/msgr-failures/osd-delay.yaml @@ -0,0 +1,9 @@ +overrides: + ceph: + conf: + global: + ms inject socket failures: 2500 + ms inject delay type: osd + ms inject delay probability: .005 + ms inject delay max: 1 + ms inject internal delays: .002 diff --git a/src/ceph/qa/suites/rados/thrash/objectstore b/src/ceph/qa/suites/rados/thrash/objectstore new file mode 120000 index 0000000..4c8ebad --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/objectstore @@ -0,0 +1 @@ +../../../objectstore
\ No newline at end of file diff --git a/src/ceph/qa/suites/rados/thrash/rados.yaml b/src/ceph/qa/suites/rados/thrash/rados.yaml new file mode 120000 index 0000000..b756e57 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/rados.yaml @@ -0,0 +1 @@ +../../../config/rados.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/rados/thrash/rocksdb.yaml b/src/ceph/qa/suites/rados/thrash/rocksdb.yaml new file mode 120000 index 0000000..f26e095 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/rocksdb.yaml @@ -0,0 +1 @@ +../../../mon_kv_backend/rocksdb.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/rados/thrash/thrashers/default.yaml b/src/ceph/qa/suites/rados/thrash/thrashers/default.yaml new file mode 100644 index 0000000..9e2b5b1 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/thrashers/default.yaml @@ -0,0 +1,17 @@ +overrides: + ceph: + log-whitelist: + - but it is still running + - objects unfound and apparently lost + conf: + osd: + osd debug reject backfill probability: .3 + osd scrub min interval: 60 + osd scrub max interval: 120 + osd max backfills: 3 + osd snap trim sleep: 2 +tasks: +- thrashosds: + timeout: 1200 + chance_pgnum_grow: 1 + chance_pgpnum_fix: 1 diff --git a/src/ceph/qa/suites/rados/thrash/thrashers/mapgap.yaml b/src/ceph/qa/suites/rados/thrash/thrashers/mapgap.yaml new file mode 100644 index 0000000..8962ff1 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/thrashers/mapgap.yaml @@ -0,0 +1,21 @@ +overrides: + ceph: + log-whitelist: + - but it is still running + - objects unfound and apparently lost + - osd_map_cache_size + conf: + mon: + mon min osdmap epochs: 2 + osd: + osd map cache size: 1 + osd scrub min interval: 60 + osd scrub max interval: 120 + osd scrub during recovery: false + osd max backfills: 6 +tasks: +- thrashosds: + timeout: 1800 + chance_pgnum_grow: 0.25 + chance_pgpnum_fix: 0.25 + chance_test_map_discontinuity: 2 diff --git a/src/ceph/qa/suites/rados/thrash/thrashers/morepggrow.yaml b/src/ceph/qa/suites/rados/thrash/thrashers/morepggrow.yaml new file mode 100644 index 0000000..91d2173 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/thrashers/morepggrow.yaml @@ -0,0 +1,22 @@ +overrides: + ceph: + conf: + osd: + osd scrub min interval: 60 + osd scrub max interval: 120 + journal throttle high multiple: 2 + journal throttle max multiple: 10 + filestore queue throttle high multiple: 2 + filestore queue throttle max multiple: 10 + osd max backfills: 9 + log-whitelist: + - but it is still running + - objects unfound and apparently lost +tasks: +- thrashosds: + timeout: 1200 + chance_pgnum_grow: 3 + chance_pgpnum_fix: 1 +openstack: +- volumes: + size: 50 diff --git a/src/ceph/qa/suites/rados/thrash/thrashers/none.yaml b/src/ceph/qa/suites/rados/thrash/thrashers/none.yaml new file mode 100644 index 0000000..e69de29 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/thrashers/none.yaml diff --git a/src/ceph/qa/suites/rados/thrash/thrashers/pggrow.yaml b/src/ceph/qa/suites/rados/thrash/thrashers/pggrow.yaml new file mode 100644 index 0000000..2a8087f --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/thrashers/pggrow.yaml @@ -0,0 +1,17 @@ +overrides: + ceph: + log-whitelist: + - but it is still running + - objects unfound and apparently lost + conf: + osd: + osd scrub min interval: 60 + osd scrub max interval: 120 + filestore odsync write: true + osd max backfills: 2 + osd snap trim sleep: .5 +tasks: +- thrashosds: + timeout: 1200 + chance_pgnum_grow: 2 + chance_pgpnum_fix: 1 diff --git a/src/ceph/qa/suites/rados/thrash/thrashosds-health.yaml b/src/ceph/qa/suites/rados/thrash/thrashosds-health.yaml new file mode 120000 index 0000000..ebf7f34 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/thrashosds-health.yaml @@ -0,0 +1 @@ +../../../tasks/thrashosds-health.yaml
\ No newline at end of file diff --git a/src/ceph/qa/suites/rados/thrash/workloads/admin_socket_objecter_requests.yaml b/src/ceph/qa/suites/rados/thrash/workloads/admin_socket_objecter_requests.yaml new file mode 100644 index 0000000..8c9764a --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/workloads/admin_socket_objecter_requests.yaml @@ -0,0 +1,13 @@ +overrides: + ceph: + conf: + client.0: + admin socket: /var/run/ceph/ceph-$name.asok +tasks: +- radosbench: + clients: [client.0] + time: 150 +- admin_socket: + client.0: + objecter_requests: + test: "http://git.ceph.com/?p={repo};a=blob_plain;f=src/test/admin_socket/objecter_requests;hb={branch}" diff --git a/src/ceph/qa/suites/rados/thrash/workloads/cache-agent-big.yaml b/src/ceph/qa/suites/rados/thrash/workloads/cache-agent-big.yaml new file mode 100644 index 0000000..0cef207 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/workloads/cache-agent-big.yaml @@ -0,0 +1,31 @@ +overrides: + ceph: + log-whitelist: + - must scrub before tier agent can activate +tasks: +- exec: + client.0: + - sudo ceph osd erasure-code-profile set myprofile crush-failure-domain=osd m=2 k=2 + - sudo ceph osd pool create base 4 4 erasure myprofile + - sudo ceph osd pool application enable base rados + - sudo ceph osd pool set base min_size 2 + - sudo ceph osd pool create cache 4 + - sudo ceph osd tier add base cache + - sudo ceph osd tier cache-mode cache writeback + - sudo ceph osd tier set-overlay base cache + - sudo ceph osd pool set cache hit_set_type bloom + - sudo ceph osd pool set cache hit_set_count 8 + - sudo ceph osd pool set cache hit_set_period 60 + - sudo ceph osd pool set cache target_max_objects 5000 +- rados: + clients: [client.0] + pools: [base] + ops: 10000 + objects: 6600 + max_seconds: 1200 + size: 1024 + op_weights: + read: 100 + write: 100 + delete: 50 + copy_from: 50 diff --git a/src/ceph/qa/suites/rados/thrash/workloads/cache-agent-small.yaml b/src/ceph/qa/suites/rados/thrash/workloads/cache-agent-small.yaml new file mode 100644 index 0000000..10d4735 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/workloads/cache-agent-small.yaml @@ -0,0 +1,30 @@ +overrides: + ceph: + crush_tunables: firefly + log-whitelist: + - must scrub before tier agent can activate +tasks: +- exec: + client.0: + - sudo ceph osd pool create base 4 + - sudo ceph osd pool application enable base rados + - sudo ceph osd pool create cache 4 + - sudo ceph osd tier add base cache + - sudo ceph osd tier cache-mode cache writeback + - sudo ceph osd tier set-overlay base cache + - sudo ceph osd pool set cache hit_set_type bloom + - sudo ceph osd pool set cache hit_set_count 8 + - sudo ceph osd pool set cache hit_set_period 60 + - sudo ceph osd pool set cache target_max_objects 250 + - sudo ceph osd pool set cache min_read_recency_for_promote 2 + - sudo ceph osd pool set cache min_write_recency_for_promote 2 +- rados: + clients: [client.0] + pools: [base] + ops: 4000 + objects: 500 + op_weights: + read: 100 + write: 100 + delete: 50 + copy_from: 50 diff --git a/src/ceph/qa/suites/rados/thrash/workloads/cache-pool-snaps-readproxy.yaml b/src/ceph/qa/suites/rados/thrash/workloads/cache-pool-snaps-readproxy.yaml new file mode 100644 index 0000000..4349743 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/workloads/cache-pool-snaps-readproxy.yaml @@ -0,0 +1,34 @@ +overrides: + ceph: + log-whitelist: + - must scrub before tier agent can activate +tasks: +- exec: + client.0: + - sudo ceph osd pool create base 4 + - sudo ceph osd pool application enable base rados + - sudo ceph osd pool create cache 4 + - sudo ceph osd tier add base cache + - sudo ceph osd tier cache-mode cache readproxy + - sudo ceph osd tier set-overlay base cache + - sudo ceph osd pool set cache hit_set_type bloom + - sudo ceph osd pool set cache hit_set_count 8 + - sudo ceph osd pool set cache hit_set_period 3600 + - sudo ceph osd pool set cache target_max_objects 250 +- rados: + clients: [client.0] + pools: [base] + ops: 4000 + objects: 500 + pool_snaps: true + op_weights: + read: 100 + write: 100 + delete: 50 + copy_from: 50 + cache_flush: 50 + cache_try_flush: 50 + cache_evict: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 diff --git a/src/ceph/qa/suites/rados/thrash/workloads/cache-pool-snaps.yaml b/src/ceph/qa/suites/rados/thrash/workloads/cache-pool-snaps.yaml new file mode 100644 index 0000000..dc3385c --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/workloads/cache-pool-snaps.yaml @@ -0,0 +1,39 @@ +overrides: + ceph: + log-whitelist: + - must scrub before tier agent can activate +tasks: +- exec: + client.0: + - sudo ceph osd pool create base 4 + - sudo ceph osd pool application enable base rados + - sudo ceph osd pool create cache 4 + - sudo ceph osd tier add base cache + - sudo ceph osd tier cache-mode cache writeback + - sudo ceph osd tier set-overlay base cache + - sudo ceph osd pool set cache hit_set_type bloom + - sudo ceph osd pool set cache hit_set_count 8 + - sudo ceph osd pool set cache hit_set_period 3600 + - sudo ceph osd pool set cache target_max_objects 250 + - sudo ceph osd pool set cache min_read_recency_for_promote 0 + - sudo ceph osd pool set cache min_write_recency_for_promote 0 +- rados: + clients: [client.0] + pools: [base] + ops: 4000 + objects: 500 + pool_snaps: true + op_weights: + read: 100 + write: 100 + delete: 50 + copy_from: 50 + cache_flush: 50 + cache_try_flush: 50 + cache_evict: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 +openstack: + - machine: + ram: 15000 # MB diff --git a/src/ceph/qa/suites/rados/thrash/workloads/cache-snaps.yaml b/src/ceph/qa/suites/rados/thrash/workloads/cache-snaps.yaml new file mode 100644 index 0000000..486d6db --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/workloads/cache-snaps.yaml @@ -0,0 +1,34 @@ +overrides: + ceph: + log-whitelist: + - must scrub before tier agent can activate +tasks: +- exec: + client.0: + - sudo ceph osd pool create base 4 + - sudo ceph osd pool application enable base rados + - sudo ceph osd pool create cache 4 + - sudo ceph osd tier add base cache + - sudo ceph osd tier cache-mode cache writeback + - sudo ceph osd tier set-overlay base cache + - sudo ceph osd pool set cache hit_set_type bloom + - sudo ceph osd pool set cache hit_set_count 8 + - sudo ceph osd pool set cache hit_set_period 3600 + - sudo ceph osd pool set cache target_max_objects 250 + - sudo ceph osd pool set cache min_read_recency_for_promote 2 +- rados: + clients: [client.0] + pools: [base] + ops: 4000 + objects: 500 + op_weights: + read: 100 + write: 100 + delete: 50 + copy_from: 50 + cache_flush: 50 + cache_try_flush: 50 + cache_evict: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 diff --git a/src/ceph/qa/suites/rados/thrash/workloads/cache.yaml b/src/ceph/qa/suites/rados/thrash/workloads/cache.yaml new file mode 100644 index 0000000..d63018f --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/workloads/cache.yaml @@ -0,0 +1,31 @@ +overrides: + ceph: + log-whitelist: + - must scrub before tier agent can activate +tasks: +- exec: + client.0: + - sudo ceph osd pool create base 4 + - sudo ceph osd pool application enable base rados + - sudo ceph osd pool create cache 4 + - sudo ceph osd tier add base cache + - sudo ceph osd tier cache-mode cache writeback + - sudo ceph osd tier set-overlay base cache + - sudo ceph osd pool set cache hit_set_type bloom + - sudo ceph osd pool set cache hit_set_count 8 + - sudo ceph osd pool set cache hit_set_period 3600 + - sudo ceph osd pool set cache min_read_recency_for_promote 0 + - sudo ceph osd pool set cache min_write_recency_for_promote 0 +- rados: + clients: [client.0] + pools: [base] + ops: 4000 + objects: 500 + op_weights: + read: 100 + write: 100 + delete: 50 + copy_from: 50 + cache_flush: 50 + cache_try_flush: 50 + cache_evict: 50 diff --git a/src/ceph/qa/suites/rados/thrash/workloads/pool-snaps-few-objects.yaml b/src/ceph/qa/suites/rados/thrash/workloads/pool-snaps-few-objects.yaml new file mode 100644 index 0000000..1f0759d --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/workloads/pool-snaps-few-objects.yaml @@ -0,0 +1,18 @@ +override: + conf: + osd: + osd deep scrub update digest min age: 0 +tasks: +- rados: + clients: [client.0] + ops: 4000 + objects: 50 + pool_snaps: true + op_weights: + read: 100 + write: 100 + delete: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 + copy_from: 50 diff --git a/src/ceph/qa/suites/rados/thrash/workloads/rados_api_tests.yaml b/src/ceph/qa/suites/rados/thrash/workloads/rados_api_tests.yaml new file mode 100644 index 0000000..23c705d --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/workloads/rados_api_tests.yaml @@ -0,0 +1,16 @@ +overrides: + ceph: + log-whitelist: + - reached quota + - \(POOL_APP_NOT_ENABLED\) + crush_tunables: hammer + conf: + client: + debug ms: 1 + debug objecter: 20 + debug rados: 20 +tasks: +- workunit: + clients: + client.0: + - rados/test.sh diff --git a/src/ceph/qa/suites/rados/thrash/workloads/radosbench.yaml b/src/ceph/qa/suites/rados/thrash/workloads/radosbench.yaml new file mode 100644 index 0000000..1b25004 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/workloads/radosbench.yaml @@ -0,0 +1,33 @@ +overrides: + ceph: + conf: + client.0: + debug ms: 1 + debug objecter: 20 + debug rados: 20 +tasks: +- full_sequential: + - radosbench: + clients: [client.0] + time: 90 + - radosbench: + clients: [client.0] + time: 90 + - radosbench: + clients: [client.0] + time: 90 + - radosbench: + clients: [client.0] + time: 90 + - radosbench: + clients: [client.0] + time: 90 + - radosbench: + clients: [client.0] + time: 90 + - radosbench: + clients: [client.0] + time: 90 + - radosbench: + clients: [client.0] + time: 90 diff --git a/src/ceph/qa/suites/rados/thrash/workloads/small-objects.yaml b/src/ceph/qa/suites/rados/thrash/workloads/small-objects.yaml new file mode 100644 index 0000000..f5a18ae --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/workloads/small-objects.yaml @@ -0,0 +1,24 @@ +overrides: + ceph: + crush_tunables: jewel + conf: + mon: + mon osd initial require min compat client: jewel +tasks: +- rados: + clients: [client.0] + ops: 400000 + max_seconds: 600 + max_in_flight: 64 + objects: 1024 + size: 16384 + op_weights: + read: 100 + write: 100 + delete: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 + copy_from: 50 + setattr: 25 + rmattr: 25 diff --git a/src/ceph/qa/suites/rados/thrash/workloads/snaps-few-objects.yaml b/src/ceph/qa/suites/rados/thrash/workloads/snaps-few-objects.yaml new file mode 100644 index 0000000..aa82d97 --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/workloads/snaps-few-objects.yaml @@ -0,0 +1,13 @@ +tasks: +- rados: + clients: [client.0] + ops: 4000 + objects: 50 + op_weights: + read: 100 + write: 100 + delete: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 + copy_from: 50 diff --git a/src/ceph/qa/suites/rados/thrash/workloads/write_fadvise_dontneed.yaml b/src/ceph/qa/suites/rados/thrash/workloads/write_fadvise_dontneed.yaml new file mode 100644 index 0000000..606dcae --- /dev/null +++ b/src/ceph/qa/suites/rados/thrash/workloads/write_fadvise_dontneed.yaml @@ -0,0 +1,8 @@ +tasks: +- rados: + clients: [client.0] + ops: 4000 + objects: 500 + write_fadvise_dontneed: true + op_weights: + write: 100 |