diff options
Diffstat (limited to 'src/ceph/qa/suites/kcephfs/recovery/tasks')
16 files changed, 205 insertions, 0 deletions
diff --git a/src/ceph/qa/suites/kcephfs/recovery/tasks/auto-repair.yaml b/src/ceph/qa/suites/kcephfs/recovery/tasks/auto-repair.yaml new file mode 100644 index 0000000..90d0e7b --- /dev/null +++ b/src/ceph/qa/suites/kcephfs/recovery/tasks/auto-repair.yaml @@ -0,0 +1,13 @@ +overrides: + ceph: + log-whitelist: + - force file system read-only + - bad backtrace + - MDS in read-only mode + - \(MDS_READ_ONLY\) + + +tasks: + - cephfs_test_runner: + modules: + - tasks.cephfs.test_auto_repair diff --git a/src/ceph/qa/suites/kcephfs/recovery/tasks/backtrace.yaml b/src/ceph/qa/suites/kcephfs/recovery/tasks/backtrace.yaml new file mode 100644 index 0000000..d740a5f --- /dev/null +++ b/src/ceph/qa/suites/kcephfs/recovery/tasks/backtrace.yaml @@ -0,0 +1,5 @@ + +tasks: + - cephfs_test_runner: + modules: + - tasks.cephfs.test_backtrace diff --git a/src/ceph/qa/suites/kcephfs/recovery/tasks/client-limits.yaml b/src/ceph/qa/suites/kcephfs/recovery/tasks/client-limits.yaml new file mode 100644 index 0000000..f816cee --- /dev/null +++ b/src/ceph/qa/suites/kcephfs/recovery/tasks/client-limits.yaml @@ -0,0 +1,20 @@ + +overrides: + ceph: + log-whitelist: + - responding to mclientcaps\(revoke\) + - not advance its oldest_client_tid + - failing to advance its oldest client/flush tid + - Too many inodes in cache + - failing to respond to cache pressure + - slow requests are blocked + - failing to respond to capability release + - MDS cache is too large + - \(MDS_CLIENT_OLDEST_TID\) + - \(MDS_CACHE_OVERSIZED\) + +tasks: + - cephfs_test_runner: + fail_on_skip: false + modules: + - tasks.cephfs.test_client_limits diff --git a/src/ceph/qa/suites/kcephfs/recovery/tasks/client-recovery.yaml b/src/ceph/qa/suites/kcephfs/recovery/tasks/client-recovery.yaml new file mode 100644 index 0000000..72ce013 --- /dev/null +++ b/src/ceph/qa/suites/kcephfs/recovery/tasks/client-recovery.yaml @@ -0,0 +1,14 @@ + +# The task interferes with the network, so we need +# to permit OSDs to complain about that. +overrides: + ceph: + log-whitelist: + - but it is still running + - slow request + - evicting unresponsive client + +tasks: + - cephfs_test_runner: + modules: + - tasks.cephfs.test_client_recovery diff --git a/src/ceph/qa/suites/kcephfs/recovery/tasks/config-commands.yaml b/src/ceph/qa/suites/kcephfs/recovery/tasks/config-commands.yaml new file mode 100644 index 0000000..cb00216 --- /dev/null +++ b/src/ceph/qa/suites/kcephfs/recovery/tasks/config-commands.yaml @@ -0,0 +1,12 @@ + +overrides: + ceph: + conf: + global: + lockdep: true + +tasks: + - cephfs_test_runner: + fail_on_skip: false + modules: + - tasks.cephfs.test_config_commands diff --git a/src/ceph/qa/suites/kcephfs/recovery/tasks/damage.yaml b/src/ceph/qa/suites/kcephfs/recovery/tasks/damage.yaml new file mode 100644 index 0000000..3f4aac9 --- /dev/null +++ b/src/ceph/qa/suites/kcephfs/recovery/tasks/damage.yaml @@ -0,0 +1,25 @@ + +overrides: + ceph: + log-whitelist: + - bad backtrace + - object missing on disk + - error reading table object + - error reading sessionmap + - Error loading MDS rank + - missing journal object + - Error recovering journal + - error decoding table object + - failed to read JournalPointer + - Corrupt directory entry + - Corrupt fnode header + - corrupt sessionmap header + - Corrupt dentry + - Scrub error on inode + - Metadata damage detected + +tasks: + - cephfs_test_runner: + modules: + - tasks.cephfs.test_damage + diff --git a/src/ceph/qa/suites/kcephfs/recovery/tasks/data-scan.yaml b/src/ceph/qa/suites/kcephfs/recovery/tasks/data-scan.yaml new file mode 100644 index 0000000..b2cd739 --- /dev/null +++ b/src/ceph/qa/suites/kcephfs/recovery/tasks/data-scan.yaml @@ -0,0 +1,18 @@ + +overrides: + ceph: + log-whitelist: + - bad backtrace + - object missing on disk + - error reading table object + - error reading sessionmap + - unmatched fragstat + - was unreadable, recreating it now + - Scrub error on inode + - Metadata damage detected + - inconsistent rstat on inode + +tasks: + - cephfs_test_runner: + modules: + - tasks.cephfs.test_data_scan diff --git a/src/ceph/qa/suites/kcephfs/recovery/tasks/failover.yaml b/src/ceph/qa/suites/kcephfs/recovery/tasks/failover.yaml new file mode 100644 index 0000000..2e4655b --- /dev/null +++ b/src/ceph/qa/suites/kcephfs/recovery/tasks/failover.yaml @@ -0,0 +1,10 @@ +overrides: + ceph: + log-whitelist: + - not responding, replacing + - \(MDS_INSUFFICIENT_STANDBY\) +tasks: + - cephfs_test_runner: + fail_on_skip: false + modules: + - tasks.cephfs.test_failover diff --git a/src/ceph/qa/suites/kcephfs/recovery/tasks/forward-scrub.yaml b/src/ceph/qa/suites/kcephfs/recovery/tasks/forward-scrub.yaml new file mode 100644 index 0000000..b92cf10 --- /dev/null +++ b/src/ceph/qa/suites/kcephfs/recovery/tasks/forward-scrub.yaml @@ -0,0 +1,14 @@ + +overrides: + ceph: + log-whitelist: + - inode wrongly marked free + - bad backtrace on inode + - inode table repaired for inode + - Scrub error on inode + - Metadata damage detected + +tasks: + - cephfs_test_runner: + modules: + - tasks.cephfs.test_forward_scrub diff --git a/src/ceph/qa/suites/kcephfs/recovery/tasks/journal-repair.yaml b/src/ceph/qa/suites/kcephfs/recovery/tasks/journal-repair.yaml new file mode 100644 index 0000000..66f819d --- /dev/null +++ b/src/ceph/qa/suites/kcephfs/recovery/tasks/journal-repair.yaml @@ -0,0 +1,14 @@ + +overrides: + ceph: + log-whitelist: + - bad backtrace on directory inode + - error reading table object + - Metadata damage detected + - slow requests are blocked + - Behind on trimming + +tasks: + - cephfs_test_runner: + modules: + - tasks.cephfs.test_journal_repair diff --git a/src/ceph/qa/suites/kcephfs/recovery/tasks/mds-flush.yaml b/src/ceph/qa/suites/kcephfs/recovery/tasks/mds-flush.yaml new file mode 100644 index 0000000..d59a8ad --- /dev/null +++ b/src/ceph/qa/suites/kcephfs/recovery/tasks/mds-flush.yaml @@ -0,0 +1,5 @@ + +tasks: + - cephfs_test_runner: + modules: + - tasks.cephfs.test_flush diff --git a/src/ceph/qa/suites/kcephfs/recovery/tasks/mds-full.yaml b/src/ceph/qa/suites/kcephfs/recovery/tasks/mds-full.yaml new file mode 100644 index 0000000..558a206 --- /dev/null +++ b/src/ceph/qa/suites/kcephfs/recovery/tasks/mds-full.yaml @@ -0,0 +1,19 @@ + +overrides: + ceph: + log-whitelist: + - OSD full dropping all updates + - OSD near full + - failsafe engaged, dropping updates + - failsafe disengaged, no longer dropping + - is full \(reached quota + conf: + osd: + osd mon report interval max: 5 + osd objectstore: memstore + memstore device bytes: 100000000 + +tasks: + - cephfs_test_runner: + modules: + - tasks.cephfs.test_full diff --git a/src/ceph/qa/suites/kcephfs/recovery/tasks/pool-perm.yaml b/src/ceph/qa/suites/kcephfs/recovery/tasks/pool-perm.yaml new file mode 100644 index 0000000..f220626 --- /dev/null +++ b/src/ceph/qa/suites/kcephfs/recovery/tasks/pool-perm.yaml @@ -0,0 +1,5 @@ + +tasks: + - cephfs_test_runner: + modules: + - tasks.cephfs.test_pool_perm diff --git a/src/ceph/qa/suites/kcephfs/recovery/tasks/sessionmap.yaml b/src/ceph/qa/suites/kcephfs/recovery/tasks/sessionmap.yaml new file mode 100644 index 0000000..9be8338 --- /dev/null +++ b/src/ceph/qa/suites/kcephfs/recovery/tasks/sessionmap.yaml @@ -0,0 +1,14 @@ + +overrides: + ceph: + conf: + global: + ms type: simple + log-whitelist: + - client session with invalid root + +tasks: + - cephfs_test_runner: + fail_on_skip: false + modules: + - tasks.cephfs.test_sessionmap diff --git a/src/ceph/qa/suites/kcephfs/recovery/tasks/strays.yaml b/src/ceph/qa/suites/kcephfs/recovery/tasks/strays.yaml new file mode 100644 index 0000000..2809fc1 --- /dev/null +++ b/src/ceph/qa/suites/kcephfs/recovery/tasks/strays.yaml @@ -0,0 +1,5 @@ + +tasks: + - cephfs_test_runner: + modules: + - tasks.cephfs.test_strays diff --git a/src/ceph/qa/suites/kcephfs/recovery/tasks/volume-client.yaml b/src/ceph/qa/suites/kcephfs/recovery/tasks/volume-client.yaml new file mode 100644 index 0000000..d738eeb --- /dev/null +++ b/src/ceph/qa/suites/kcephfs/recovery/tasks/volume-client.yaml @@ -0,0 +1,12 @@ + +overrides: + ceph: + conf: + global: + ms type: simple + +tasks: + - cephfs_test_runner: + fail_on_skip: false + modules: + - tasks.cephfs.test_volume_client |