blob: 6c92fb0b6e1174fb1c0636c21ca68ed975269851 (
plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
|
#!/bin/bash
# Copyright 2017 AT&T Intellectual Property, Inc
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
#. What this is: script to setup a Ceph-based SDS (Software Defined Storage)
#. service for a kubernetes cluster, using Helm as deployment tool.
#. Prerequisites:
#. - Ubuntu xenial server for master and agent nodes
#. - key-based auth setup for ssh/scp between master and agent nodes
#. - 192.168.0.0/16 should not be used on your server network interface subnets
#. Usage:
# Intended to be called from k8s-cluster.sh in this folder. To run directly:
#. $ bash ceph-helm.sh "<nodes>" <cluster-net> <public-net> [ceph_dev]
#. nodes: space-separated list of ceph node IPs
#. cluster-net: CIDR of ceph cluster network e.g. 10.0.0.1/24
#. public-net: CIDR of public network
#. ceph_dev: disk to use for ceph. ***MUST NOT BE USED FOR ANY OTHER PURPOSE***
#. if not provided, ceph data will be stored on osd nodes in /ceph
#.
#. Status: work in progress, incomplete
#
function setup_ceph() {
nodes=$1
private_net=$2
public_net=$3
dev=$4
# per https://github.com/att/netarbiter/tree/master/sds/ceph-docker/examples/helm
echo "${FUNCNAME[0]}: Clone netarbiter"
git clone https://github.com/att/netarbiter.git
echo "${FUNCNAME[0]}: Create a .kube/config secret so that a K8s job could run kubectl inside the container"
cd netarbiter/sds/ceph-docker/examples/helm
kubectl create namespace ceph
./create-secret-kube-config.sh ceph
./helm-install-ceph.sh my_ceph_chart $private_net $public_net
kubedns=$(kubectl get service -o json --namespace kube-system kube-dns | \
jq -r '.spec.clusterIP')
for node in $nodes; do
echo "${FUNCNAME[0]}: setup resolv.conf for $node"
echo <<EOF | sudo tee -a /etc/resolv.conf
nameserver $kubedns
search ceph.svc.cluster.local svc.cluster.local cluster.local
EOF
echo "${FUNCNAME[0]}: Zap disk $dev at $node"
ssh -x -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no \
ubuntu@$node ceph-disk zap $dev
echo "${FUNCNAME[0]}: Run ceph-osd at $node"
name=$(ssh -x -o UserKnownHostsFile=/dev/null -o StrictHostKeyChecking=no \
ubuntu@$node hostname)
./helm-install-ceph-osd.sh $name /dev/$dev
done
echo "${FUNCNAME[0]}: Activate Ceph for namespace 'default'"
./activate-namespace.sh default
echo "${FUNCNAME[0]}: Relax access control rules"
kubectl replace -f relax-rbac-k8s1.7.yaml
# TODO: verification tests
}
if [[ "$1" != "" ]]; then
setup_ceph "$1" $2 $3 $4
else
grep '#. ' $0
fi
|