#ceph osd tree # id weight type name up/down reweight -1 76.47 root default -2 32.72 host ceph-s-01 0 7.27 osd.0 up 1 1 7.27 osd.1 up 1 2 9.09 osd.2 up 1 3 9.09 osd.3 up 1 -3 43.75 host ceph-s-02 4 10.91 osd.4 up 1 5 0.11 osd.5 up 1 6 10.91 osd.6 up 1 7 10.91 osd.7 up 1 8 10.91 osd.8 up 1
On 08.05.2014 19:11, Craig Lewis wrote:
What does `ceph osd tree` output? On 5/8/14 07:30 , Georg Höllrigl wrote:Hello, We've a fresh cluster setup - with Ubuntu 14.04 and ceph firefly. By now I've tried this multiple times - but the result keeps the same and shows me lots of troubles (the cluster is empty, no client has accessed it) #ceph -s cluster b04fc583-9e71-48b7-a741-92f4dff4cfef health HEALTH_WARN 470 pgs stale; 470 pgs stuck stale; 18 pgs stuck unclean; 26 requests are blocked > 32 sec monmap e2: 3 mons at {ceph-m-01=10.0.0.100:6789/0,ceph-m-02=10.0.1.101:6789/0,ceph-m-03=10.0.1.102:6789/0}, election epoch 8, quorum 0,1,2 ceph-m-01,ceph-m-02,ceph-m-03 osdmap e409: 9 osds: 9 up, 9 in pgmap v1231: 480 pgs, 9 pools, 822 bytes data, 43 objects 9373 MB used, 78317 GB / 78326 GB avail 451 stale+active+clean 1 stale+active+clean+scrubbing 10 active+clean 18 stale+active+remapped
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
