summaryrefslogtreecommitdiffstats
path: root/src/ceph/qa/standalone/crush/crush-classes.sh
blob: bcaab3fa0e8898093f137bad3e3e77a3ff0e478d (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
#!/bin/bash
#
# Copyright (C) 2017 Red Hat <contact@redhat.com>
#
# Author: Loic Dachary <loic@dachary.org>
#
# This program is free software; you can redistribute it and/or modify
# it under the terms of the GNU Library Public License as published by
# the Free Software Foundation; either version 2, or (at your option)
# any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
# GNU Library Public License for more details.
#

source $CEPH_ROOT/qa/standalone/ceph-helpers.sh

function run() {
    local dir=$1
    shift

    export CEPH_MON="127.0.0.1:7130" # git grep '\<7130\>' : there must be only one
    export CEPH_ARGS
    CEPH_ARGS+="--fsid=$(uuidgen) --auth-supported=none "
    CEPH_ARGS+="--mon-host=$CEPH_MON "
    #
    # Disable auto-class, so we can inject device class manually below
    #
    CEPH_ARGS+="--osd-class-update-on-start=false "

    local funcs=${@:-$(set | sed -n -e 's/^\(TEST_[0-9a-z_]*\) .*/\1/p')}
    for func in $funcs ; do
        setup $dir || return 1
        $func $dir || return 1
        teardown $dir || return 1
    done
}

function add_something() {
    local dir=$1
    local obj=${2:-SOMETHING}

    local payload=ABCDEF
    echo $payload > $dir/ORIGINAL
    rados --pool rbd put $obj $dir/ORIGINAL || return 1
}

function get_osds_up() {
    local poolname=$1
    local objectname=$2

    local osds=$(ceph --format xml osd map $poolname $objectname 2>/dev/null | \
        $XMLSTARLET sel -t -m "//up/osd" -v . -o ' ')
    # get rid of the trailing space
    echo $osds
}

function TEST_classes() {
    local dir=$1

    run_mon $dir a || return 1
    run_osd $dir 0 || return 1
    run_osd $dir 1 || return 1
    run_osd $dir 2 || return 1
    create_rbd_pool || return 1

    test "$(get_osds_up rbd SOMETHING)" == "1 2 0" || return 1
    add_something $dir SOMETHING || return 1

    #
    # osd.0 has class ssd and the rule is modified
    # to only take ssd devices.
    #
    ceph osd getcrushmap > $dir/map || return 1
    crushtool -d $dir/map -o $dir/map.txt || return 1
    ${SED} -i \
        -e '/device 0 osd.0/s/$/ class ssd/' \
        -e '/step take default/s/$/ class ssd/' \
        $dir/map.txt || return 1
    crushtool -c $dir/map.txt -o $dir/map-new || return 1
    ceph osd setcrushmap -i $dir/map-new || return 1

    #
    # There can only be one mapping since there only is
    # one device with ssd class.
    #
    ok=false
    for delay in 2 4 8 16 32 64 128 256 ; do
        if test "$(get_osds_up rbd SOMETHING_ELSE)" == "0" ; then
            ok=true
            break
        fi
        sleep $delay
        ceph osd dump # for debugging purposes
        ceph pg dump # for debugging purposes
    done
    $ok || return 1
    #
    # Writing keeps working because the pool is min_size 1 by
    # default.
    #
    add_something $dir SOMETHING_ELSE || return 1

    #
    # Sanity check that the rule indeed has ssd
    # generated bucket with a name including ~ssd.
    #
    ceph osd crush dump | grep -q '~ssd' || return 1
}

function TEST_set_device_class() {
    local dir=$1

    TEST_classes $dir || return 1

    ceph osd crush set-device-class ssd osd.0 || return 1
    ceph osd crush class ls-osd ssd | grep 0 || return 1
    ceph osd crush set-device-class ssd osd.1 || return 1
    ceph osd crush class ls-osd ssd | grep 1 || return 1
    ceph osd crush set-device-class ssd 0 1 || return 1 # should be idempotent

    ok=false
    for delay in 2 4 8 16 32 64 128 256 ; do
        if test "$(get_osds_up rbd SOMETHING_ELSE)" == "0 1" ; then
            ok=true
            break
        fi
        sleep $delay
        ceph osd crush dump
        ceph osd dump # for debugging purposes
        ceph pg dump # for debugging purposes
    done
    $ok || return 1
}

function TEST_mon_classes() {
    local dir=$1

    run_mon $dir a || return 1
    run_osd $dir 0 || return 1
    run_osd $dir 1 || return 1
    run_osd $dir 2 || return 1
    create_rbd_pool || return 1

    test "$(get_osds_up rbd SOMETHING)" == "1 2 0" || return 1
    add_something $dir SOMETHING || return 1

    # test rm-device-class
    ceph osd crush set-device-class aaa osd.0 || return 1
    ceph osd tree | grep -q 'aaa' || return 1
    ceph osd crush dump | grep -q '~aaa' || return 1
    ceph osd crush tree --show-shadow | grep -q '~aaa' || return 1
    ceph osd crush set-device-class bbb osd.1 || return 1
    ceph osd tree | grep -q 'bbb' || return 1
    ceph osd crush dump | grep -q '~bbb' || return 1
    ceph osd crush tree --show-shadow | grep -q '~bbb' || return 1
    ceph osd crush set-device-class ccc osd.2 || return 1
    ceph osd tree | grep -q 'ccc' || return 1
    ceph osd crush dump | grep -q '~ccc' || return 1
    ceph osd crush tree --show-shadow | grep -q '~ccc' || return 1
    ceph osd crush rm-device-class 0 || return 1
    ceph osd tree | grep -q 'aaa' && return 1
    ceph osd crush class ls | grep -q 'aaa' && return 1 # class 'aaa' should gone
    ceph osd crush rm-device-class 1 || return 1
    ceph osd tree | grep -q 'bbb' && return 1
    ceph osd crush class ls | grep -q 'bbb' && return 1 # class 'bbb' should gone
    ceph osd crush rm-device-class 2 || return 1
    ceph osd tree | grep -q 'ccc' && return 1
    ceph osd crush class ls | grep -q 'ccc' && return 1 # class 'ccc' should gone
    ceph osd crush set-device-class asdf all || return 1
    ceph osd tree | grep -q 'asdf' || return 1
    ceph osd crush dump | grep -q '~asdf' || return 1
    ceph osd crush tree --show-shadow | grep -q '~asdf' || return 1
    ceph osd crush rule create-replicated asdf-rule default host asdf || return 1
    ceph osd crush rm-device-class all || return 1
    ceph osd tree | grep -q 'asdf' && return 1
    ceph osd crush class ls | grep -q 'asdf' || return 1 # still referenced by asdf-rule

    ceph osd crush set-device-class abc osd.2 || return 1
    ceph osd crush move osd.2 root=foo rack=foo-rack host=foo-host || return 1
    out=`ceph osd tree |awk '$1 == 2 && $2 == "abc" {print $0}'`
    if [ "$out" == "" ]; then
        return 1
    fi

    # verify 'crush move' too
    ceph osd crush dump | grep -q 'foo~abc' || return 1
    ceph osd crush tree --show-shadow | grep -q 'foo~abc' || return 1
    ceph osd crush dump | grep -q 'foo-rack~abc' || return 1
    ceph osd crush tree --show-shadow | grep -q 'foo-rack~abc' || return 1
    ceph osd crush dump | grep -q 'foo-host~abc' || return 1
    ceph osd crush tree --show-shadow | grep -q 'foo-host~abc' || return 1
    ceph osd crush rm-device-class osd.2 || return 1
    # restore class, so we can continue to test create-replicated
    ceph osd crush set-device-class abc osd.2 || return 1

    ceph osd crush rule create-replicated foo-rule foo host abc || return 1

    # test set-device-class implicitly change class
    ceph osd crush set-device-class hdd osd.0 || return 1
    expect_failure $dir EBUSY ceph osd crush set-device-class nvme osd.0 || return 1

    # test class rename
    ceph osd crush rm-device-class all || return 1
    ceph osd crush set-device-class class_1 all || return 1
    ceph osd crush class ls | grep 'class_1' || return 1
    ceph osd crush tree --show-shadow | grep 'class_1' || return 1
    ceph osd crush rule create-replicated class_1_rule default host class_1 || return 1
    ceph osd crush class rename class_1 class_2
    ceph osd crush class rename class_1 class_2 # idempotent
    ceph osd crush class ls | grep 'class_1' && return 1
    ceph osd crush tree --show-shadow | grep 'class_1' && return 1
    ceph osd crush class ls | grep 'class_2' || return 1
    ceph osd crush tree --show-shadow | grep 'class_2' || return 1
}

main crush-classes "$@"

# Local Variables:
# compile-command: "cd ../../../build ; ln -sf ../src/ceph-disk/ceph_disk/main.py bin/ceph-disk && make -j4 && ../src/test/crush/crush-classes.sh"
# End: