Hi all,

I had a question, I'm in the middle of a new ceph deploy cluster and I've 6
OSD servers between two racks, so rack1 would have osdserver1,3 and 5, and
rack2 osdserver2,4 and 6. I've edited the following crush map and I want to
know if it's ok and also if the objects would be stored one on each
rack-host. So, if I lost one rack, I had one copy on the other rack/server:

*http://pastebin.com/raw/QJf1VeeJ <http://pastebin.com/raw/QJf1VeeJ>*

Also If I need to run any command in order to 'apply' the new crush map to
the existing pools (actually only two):

- 0 rbd            (pg_num: 4096 | pgp_num: 4096 | size: 2 | min_size: 1)
- 1 cinder-volumes (pg_num: 4096 | pgp_num: 4096 | size: 2 | min_size: 1)

# ceph --cluster cephIB osd tree
ID WEIGHT   TYPE NAME            UP/DOWN REWEIGHT PRIMARY-AFFINITY
-1 51.29668 root default
-8 26.00958     rack cage5-rack1
-2  8.66986         host cibn01
 0  0.72249             osd.0         up  1.00000          1.00000
 1  0.72249             osd.1         up  1.00000          1.00000
 2  0.72249             osd.2         up  1.00000          1.00000
 3  0.72249             osd.3         up  1.00000          1.00000
 4  0.72249             osd.4         up  1.00000          1.00000
 5  0.72249             osd.5         up  1.00000          1.00000
 6  0.72249             osd.6         up  1.00000          1.00000
 7  0.72249             osd.7         up  1.00000          1.00000
 8  0.72249             osd.8         up  1.00000          1.00000
 9  0.72249             osd.9         up  1.00000          1.00000
10  0.72249             osd.10        up  1.00000          1.00000
11  0.72249             osd.11        up  1.00000          1.00000
-4  8.66986         host cibn03
24  0.72249             osd.24        up  1.00000          1.00000
25  0.72249             osd.25        up  1.00000          1.00000
26  0.72249             osd.26        up  1.00000          1.00000
27  0.72249             osd.27        up  1.00000          1.00000
28  0.72249             osd.28        up  1.00000          1.00000
29  0.72249             osd.29        up  1.00000          1.00000
30  0.72249             osd.30        up  1.00000          1.00000
31  0.72249             osd.31        up  1.00000          1.00000
32  0.72249             osd.32        up  1.00000          1.00000
33  0.72249             osd.33        up  1.00000          1.00000
34  0.72249             osd.34        up  1.00000          1.00000
35  0.72249             osd.35        up  1.00000          1.00000
-6  8.66986         host cibn05
48  0.72249             osd.48        up  1.00000          1.00000
49  0.72249             osd.49        up  1.00000          1.00000
50  0.72249             osd.50        up  1.00000          1.00000
51  0.72249             osd.51        up  1.00000          1.00000
52  0.72249             osd.52        up  1.00000          1.00000
53  0.72249             osd.53        up  1.00000          1.00000
54  0.72249             osd.54        up  1.00000          1.00000
55  0.72249             osd.55        up  1.00000          1.00000
56  0.72249             osd.56        up  1.00000          1.00000
57  0.72249             osd.57        up  1.00000          1.00000
58  0.72249             osd.58        up  1.00000          1.00000
59  0.72249             osd.59        up  1.00000          1.00000
-9 25.28709     rack cage5-rack2
-3  8.66986         host cibn02
12  0.72249             osd.12        up  1.00000          1.00000
13  0.72249             osd.13        up  1.00000          1.00000
14  0.72249             osd.14        up  1.00000          1.00000
15  0.72249             osd.15        up  1.00000          1.00000
16  0.72249             osd.16        up  1.00000          1.00000
17  0.72249             osd.17        up  1.00000          1.00000
18  0.72249             osd.18        up  1.00000          1.00000
19  0.72249             osd.19        up  1.00000          1.00000
20  0.72249             osd.20        up  1.00000          1.00000
21  0.72249             osd.21        up  1.00000          1.00000
22  0.72249             osd.22        up  1.00000          1.00000
23  0.72249             osd.23        up  1.00000          1.00000
-5  8.66986         host cibn04
36  0.72249             osd.36        up  1.00000          1.00000
37  0.72249             osd.37        up  1.00000          1.00000
38  0.72249             osd.38        up  1.00000          1.00000
39  0.72249             osd.39        up  1.00000          1.00000
40  0.72249             osd.40        up  1.00000          1.00000
41  0.72249             osd.41        up  1.00000          1.00000
42  0.72249             osd.42        up  1.00000          1.00000
43  0.72249             osd.43        up  1.00000          1.00000
44  0.72249             osd.44        up  1.00000          1.00000
45  0.72249             osd.45        up  1.00000          1.00000
46  0.72249             osd.46        up  1.00000          1.00000
47  0.72249             osd.47        up  1.00000          1.00000
-7  7.94737         host cibn06
60  0.72249             osd.60        up  1.00000          1.00000
61  0.72249             osd.61        up  1.00000          1.00000
62  0.72249             osd.62        up  1.00000          1.00000
63  0.72249             osd.63        up  1.00000          1.00000
64  0.72249             osd.64        up  1.00000          1.00000
65  0.72249             osd.65        up  1.00000          1.00000
66  0.72249             osd.66        up  1.00000          1.00000
67  0.72249             osd.67        up  1.00000          1.00000
68  0.72249             osd.68        up  1.00000          1.00000
69  0.72249             osd.69        up  1.00000          1.00000
70  0.72249             osd.70        up  1.00000          1.00000


Ceph version 9.2.1 (752b6a3020c3de74e07d2a8b4c5e48dab5a6b6fd)


Thanks in advance,

Best,

*German*
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to