Hi all, I had a question, I'm in the middle of a new ceph deploy cluster and I've 6 OSD servers between two racks, so rack1 would have osdserver1,3 and 5, and rack2 osdserver2,4 and 6. I've edited the following crush map and I want to know if it's ok and also if the objects would be stored one on each rack-host. So, if I lost one rack, I had one copy on the other rack/server:
*http://pastebin.com/raw/QJf1VeeJ <http://pastebin.com/raw/QJf1VeeJ>* Also If I need to run any command in order to 'apply' the new crush map to the existing pools (actually only two): - 0 rbd (pg_num: 4096 | pgp_num: 4096 | size: 2 | min_size: 1) - 1 cinder-volumes (pg_num: 4096 | pgp_num: 4096 | size: 2 | min_size: 1) # ceph --cluster cephIB osd tree ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY -1 51.29668 root default -8 26.00958 rack cage5-rack1 -2 8.66986 host cibn01 0 0.72249 osd.0 up 1.00000 1.00000 1 0.72249 osd.1 up 1.00000 1.00000 2 0.72249 osd.2 up 1.00000 1.00000 3 0.72249 osd.3 up 1.00000 1.00000 4 0.72249 osd.4 up 1.00000 1.00000 5 0.72249 osd.5 up 1.00000 1.00000 6 0.72249 osd.6 up 1.00000 1.00000 7 0.72249 osd.7 up 1.00000 1.00000 8 0.72249 osd.8 up 1.00000 1.00000 9 0.72249 osd.9 up 1.00000 1.00000 10 0.72249 osd.10 up 1.00000 1.00000 11 0.72249 osd.11 up 1.00000 1.00000 -4 8.66986 host cibn03 24 0.72249 osd.24 up 1.00000 1.00000 25 0.72249 osd.25 up 1.00000 1.00000 26 0.72249 osd.26 up 1.00000 1.00000 27 0.72249 osd.27 up 1.00000 1.00000 28 0.72249 osd.28 up 1.00000 1.00000 29 0.72249 osd.29 up 1.00000 1.00000 30 0.72249 osd.30 up 1.00000 1.00000 31 0.72249 osd.31 up 1.00000 1.00000 32 0.72249 osd.32 up 1.00000 1.00000 33 0.72249 osd.33 up 1.00000 1.00000 34 0.72249 osd.34 up 1.00000 1.00000 35 0.72249 osd.35 up 1.00000 1.00000 -6 8.66986 host cibn05 48 0.72249 osd.48 up 1.00000 1.00000 49 0.72249 osd.49 up 1.00000 1.00000 50 0.72249 osd.50 up 1.00000 1.00000 51 0.72249 osd.51 up 1.00000 1.00000 52 0.72249 osd.52 up 1.00000 1.00000 53 0.72249 osd.53 up 1.00000 1.00000 54 0.72249 osd.54 up 1.00000 1.00000 55 0.72249 osd.55 up 1.00000 1.00000 56 0.72249 osd.56 up 1.00000 1.00000 57 0.72249 osd.57 up 1.00000 1.00000 58 0.72249 osd.58 up 1.00000 1.00000 59 0.72249 osd.59 up 1.00000 1.00000 -9 25.28709 rack cage5-rack2 -3 8.66986 host cibn02 12 0.72249 osd.12 up 1.00000 1.00000 13 0.72249 osd.13 up 1.00000 1.00000 14 0.72249 osd.14 up 1.00000 1.00000 15 0.72249 osd.15 up 1.00000 1.00000 16 0.72249 osd.16 up 1.00000 1.00000 17 0.72249 osd.17 up 1.00000 1.00000 18 0.72249 osd.18 up 1.00000 1.00000 19 0.72249 osd.19 up 1.00000 1.00000 20 0.72249 osd.20 up 1.00000 1.00000 21 0.72249 osd.21 up 1.00000 1.00000 22 0.72249 osd.22 up 1.00000 1.00000 23 0.72249 osd.23 up 1.00000 1.00000 -5 8.66986 host cibn04 36 0.72249 osd.36 up 1.00000 1.00000 37 0.72249 osd.37 up 1.00000 1.00000 38 0.72249 osd.38 up 1.00000 1.00000 39 0.72249 osd.39 up 1.00000 1.00000 40 0.72249 osd.40 up 1.00000 1.00000 41 0.72249 osd.41 up 1.00000 1.00000 42 0.72249 osd.42 up 1.00000 1.00000 43 0.72249 osd.43 up 1.00000 1.00000 44 0.72249 osd.44 up 1.00000 1.00000 45 0.72249 osd.45 up 1.00000 1.00000 46 0.72249 osd.46 up 1.00000 1.00000 47 0.72249 osd.47 up 1.00000 1.00000 -7 7.94737 host cibn06 60 0.72249 osd.60 up 1.00000 1.00000 61 0.72249 osd.61 up 1.00000 1.00000 62 0.72249 osd.62 up 1.00000 1.00000 63 0.72249 osd.63 up 1.00000 1.00000 64 0.72249 osd.64 up 1.00000 1.00000 65 0.72249 osd.65 up 1.00000 1.00000 66 0.72249 osd.66 up 1.00000 1.00000 67 0.72249 osd.67 up 1.00000 1.00000 68 0.72249 osd.68 up 1.00000 1.00000 69 0.72249 osd.69 up 1.00000 1.00000 70 0.72249 osd.70 up 1.00000 1.00000 Ceph version 9.2.1 (752b6a3020c3de74e07d2a8b4c5e48dab5a6b6fd) Thanks in advance, Best, *German*
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com