Hi,
Do you have by any chance disabled automatic crushmap updates in your
ceph config?
osd crush update on start = false
If this is the case, and you move disks around hosts, they won't update
their position/host in the crushmap, even if the crushmap does not
reflect reality.
Regards,
Mart
On 01/08/2016 02:16 AM, Wade Holler wrote:
> Sure. Apologies for all the text: We have 12 Nodes for OSDs, 15 OSDs
> per node, but I will only include a sample:
>
> ceph osd tree | head -35
>
> ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY
>
> -1 130.98450 root default
>
> -2 5.82153 host cpn00001
>
> 4 0.72769 osd.4 up 1.00000 1.00000
>
> 14 0.72769 osd.14 up 1.00000 1.00000
>
> 3 0.72769 osd.3 up 1.00000 1.00000
>
> 24 0.72769 osd.24 up 1.00000 1.00000
>
> 5 0.72769 osd.5 up 1.00000 1.00000
>
> 2 0.72769 osd.2 up 1.00000 1.00000
>
> 17 0.72769 osd.17 up 1.00000 1.00000
>
> 69 0.72769 osd.69 up 1.00000 1.00000
>
> -3 6.54922 host cpn00003
>
> 7 0.72769 osd.7 up 1.00000 1.00000
>
> 8 0.72769 osd.8 up 1.00000 1.00000
>
> 9 0.72769 osd.9 up 1.00000 1.00000
>
> 0 0.72769 osd.0 up 1.00000 1.00000
>
> 28 0.72769 osd.28 up 1.00000 1.00000
>
> 10 0.72769 osd.10 up 1.00000 1.00000
>
> 1 0.72769 osd.1 up 1.00000 1.00000
>
> 6 0.72769 osd.6 up 1.00000 1.00000
>
> 29 0.72769 osd.29 up 1.00000 1.00000
>
> -4 2.91077 host cpn00004
>
>
> Compared with the actual processes that are running:
>
>
> [root@cpx00001 ~]# ssh cpn00001 ps -ef | grep ceph\-osd
>
> ceph 92638 1 26 16:19 ? 01:00:55
> /usr/bin/ceph-osd-f --cluster ceph --id 6 --setuser ceph --setgroup ceph
>
> ceph 92667 1 20 16:19 ? 00:48:04
> /usr/bin/ceph-osd-f --cluster ceph --id 0 --setuser ceph --setgroup ceph
>
> ceph 92673 1 18 16:19 ? 00:42:48
> /usr/bin/ceph-osd-f --cluster ceph --id 8 --setuser ceph --setgroup ceph
>
> ceph 92681 1 19 16:19 ? 00:45:52
> /usr/bin/ceph-osd-f --cluster ceph --id 7 --setuser ceph --setgroup ceph
>
> ceph 92701 1 15 16:19 ? 00:36:05
> /usr/bin/ceph-osd-f --cluster ceph --id 12 --setuser ceph --setgroup ceph
>
> ceph 92748 1 14 16:19 ? 00:34:07
> /usr/bin/ceph-osd-f --cluster ceph --id 10 --setuser ceph --setgroup ceph
>
> ceph 92756 1 16 16:19 ? 00:38:40
> /usr/bin/ceph-osd-f --cluster ceph --id 9 --setuser ceph --setgroup ceph
>
> ceph 92758 1 17 16:19 ? 00:39:28
> /usr/bin/ceph-osd-f --cluster ceph --id 13 --setuser ceph --setgroup ceph
>
> ceph 92777 1 19 16:19 ? 00:46:17
> /usr/bin/ceph-osd-f --cluster ceph --id 1 --setuser ceph --setgroup ceph
>
> ceph 92988 1 18 16:19 ? 00:42:47
> /usr/bin/ceph-osd-f --cluster ceph --id 5 --setuser ceph --setgroup ceph
>
> ceph 93058 1 18 16:19 ? 00:43:18
> /usr/bin/ceph-osd-f --cluster ceph --id 11 --setuser ceph --setgroup ceph
>
> ceph 93078 1 17 16:19 ? 00:41:38
> /usr/bin/ceph-osd-f --cluster ceph --id 14 --setuser ceph --setgroup ceph
>
> ceph 93127 1 15 16:19 ? 00:36:29
> /usr/bin/ceph-osd-f --cluster ceph --id 4 --setuser ceph --setgroup ceph
>
> ceph 93130 1 17 16:19 ? 00:40:44
> /usr/bin/ceph-osd-f --cluster ceph --id 2 --setuser ceph --setgroup ceph
>
> ceph 93173 1 21 16:19 ? 00:49:37
> /usr/bin/ceph-osd-f --cluster ceph --id 3 --setuser ceph --setgroup ceph
>
> [root@cpx00001 ~]# ssh cpn00003 ps -ef | grep ceph\-osd
>
> ceph 82454 1 18 16:19 ? 00:43:58
> /usr/bin/ceph-osd-f --cluster ceph --id 25 --setuser ceph --setgroup ceph
>
> ceph 82464 1 24 16:19 ? 00:55:40
> /usr/bin/ceph-osd-f --cluster ceph --id 21 --setuser ceph --setgroup ceph
>
> ceph 82473 1 21 16:19 ? 00:50:14
> /usr/bin/ceph-osd-f --cluster ceph --id 17 --setuser ceph --setgroup ceph
>
> ceph 82612 1 19 16:19 ? 00:45:25
> /usr/bin/ceph-osd-f --cluster ceph --id 22 --setuser ceph --setgroup ceph
>
> ceph 82629 1 20 16:19 ? 00:48:38
> /usr/bin/ceph-osd-f --cluster ceph --id 16 --setuser ceph --setgroup ceph
>
> ceph 82651 1 16 16:19 ? 00:39:24
> /usr/bin/ceph-osd-f --cluster ceph --id 20 --setuser ceph --setgroup ceph
>
> ceph 82687 1 17 16:19 ? 00:40:31
> /usr/bin/ceph-osd-f --cluster ceph --id 18 --setuser ceph --setgroup ceph
>
> ceph 82697 1 26 16:19 ? 01:02:12
> /usr/bin/ceph-osd-f --cluster ceph --id 23 --setuser ceph --setgroup ceph
>
> ceph 82719 1 20 16:19 ? 00:47:15
> /usr/bin/ceph-osd-f --cluster ceph --id 15 --setuser ceph --setgroup ceph
>
> ceph 82722 1 14 16:19 ? 00:33:41
> /usr/bin/ceph-osd-f --cluster ceph --id 28 --setuser ceph --setgroup ceph
>
> ceph 82725 1 14 16:19 ? 00:33:16
> /usr/bin/ceph-osd-f --cluster ceph --id 26 --setuser ceph --setgroup ceph
>
> ceph 82743 1 14 16:19 ? 00:34:17
> /usr/bin/ceph-osd-f --cluster ceph --id 29 --setuser ceph --setgroup ceph
>
> ceph 82769 1 19 16:19 ? 00:46:00
> /usr/bin/ceph-osd-f --cluster ceph --id 19 --setuser ceph --setgroup ceph
>
> ceph 82816 1 13 16:19 ? 00:30:26
> /usr/bin/ceph-osd-f --cluster ceph --id 27 --setuser ceph --setgroup ceph
>
> ceph 82828 1 27 16:19 ? 01:04:38
> /usr/bin/ceph-osd-f --cluster ceph --id 24 --setuser ceph --setgroup ceph
>
> [root@cpx00001 ~]#
>
>
> Looks like the crushmap is bad also:
>
> (Cluster appears to be operating ok but this really concerns me.)
>
> # begin crush map
>
> tunable choose_local_tries 0
>
> tunable choose_local_fallback_tries 0
>
> tunable choose_total_tries 50
>
> tunable chooseleaf_descend_once 1
>
> tunable straw_calc_version 1
>
>
> # devices
>
> device 0 osd.0
>
> device 1 osd.1
>
> device 2 osd.2
>
> device 3 osd.3
>
> device 4 osd.4
>
> device 5 osd.5
>
> device 6 osd.6
>
> device 7 osd.7
>
> device 8 osd.8
>
> device 9 osd.9
>
> device 10 osd.10
>
> device 11 osd.11
>
> device 12 osd.12
>
> device 13 osd.13
>
> device 14 osd.14
>
> device 15 osd.15
>
> device 16 osd.16
>
> device 17 osd.17
>
> device 18 osd.18
>
> device 19 osd.19
>
> device 20 osd.20
>
> device 21 osd.21
>
> device 22 osd.22
>
> device 23 osd.23
>
> device 24 osd.24
>
> device 25 osd.25
>
> device 26 osd.26
>
> device 27 osd.27
>
> device 28 osd.28
>
> device 29 osd.29
>
> device 30 osd.30
>
> device 31 osd.31
>
> device 32 osd.32
>
> device 33 osd.33
>
> device 34 osd.34
>
> device 35 osd.35
>
> device 36 osd.36
>
> device 37 osd.37
>
> ...
>
>
> # types
>
> type 0 osd
>
> type 1 host
>
> type 2 chassis
>
> type 3 rack
>
> type 4 row
>
> type 5 pdu
>
> type 6 pod
>
> type 7 room
>
> type 8 datacenter
>
> type 9 region
>
> type 10 root
>
>
> # buckets
>
> host cpn00001 {
>
> id -2 # do not change unnecessarily
>
> # weight 5.822
>
> alg straw
>
> hash 0 # rjenkins1
>
> item osd.4 weight 0.728
>
> item osd.14 weight 0.728
>
> item osd.3 weight 0.728
>
> item osd.24 weight 0.728
>
> item osd.5 weight 0.728
>
> item osd.2 weight 0.728
>
> item osd.17 weight 0.728
>
> item osd.69 weight 0.728
>
> }
>
> host cpn00003 {
>
> id -3 # do not change unnecessarily
>
> # weight 6.549
>
> alg straw
>
> hash 0 # rjenkins1
>
> item osd.7 weight 0.728
>
> item osd.8 weight 0.728
>
> item osd.9 weight 0.728
>
> item osd.0 weight 0.728
>
> item osd.28 weight 0.728
>
> item osd.10 weight 0.728
>
> item osd.1 weight 0.728
>
> item osd.6 weight 0.728
>
> item osd.29 weight 0.728
>
> }
>
> host cpn00004 {....
>
>
>
> Thank you for your review !!!!!
>
> Wade
>
>
>
>
>
> On Thu, Jan 7, 2016 at 6:03 PM Shinobu Kinjo <[email protected]
> <mailto:[email protected]>> wrote:
>
> Can you share the output with us?
>
> Rgds,
> Shinobu
>
> ----- Original Message -----
> From: "Wade Holler" <[email protected]
> <mailto:[email protected]>>
> To: "ceph-users" <[email protected]
> <mailto:[email protected]>>
> Sent: Friday, January 8, 2016 7:29:07 AM
> Subject: [ceph-users] ceph osd tree output
>
> Sometimes my ceph osd tree output is wrong. Ie. Wrong osds on the
> wrong hosts ?
>
> Anyone else have this issue?
>
> I have seen this at Infernalis and Jewell.
>
> Thanks
> Wade
>
> _______________________________________________
> ceph-users mailing list
> [email protected] <mailto:[email protected]>
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
signature.asc
Description: OpenPGP digital signature
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
