#ceph osd tree
# id    weight  type name       up/down reweight
-1      76.47   root default
-2      32.72           host ceph-s-01
0       7.27                    osd.0   up      1
1       7.27                    osd.1   up      1
2       9.09                    osd.2   up      1
3       9.09                    osd.3   up      1
-3      43.75           host ceph-s-02
4       10.91                   osd.4   up      1
5       0.11                    osd.5   up      1
6       10.91                   osd.6   up      1
7       10.91                   osd.7   up      1
8       10.91                   osd.8   up      1


On 08.05.2014 19:11, Craig Lewis wrote:
What does `ceph osd tree` output?

On 5/8/14 07:30 , Georg Höllrigl wrote:
Hello,

We've a fresh cluster setup - with Ubuntu 14.04 and ceph firefly. By
now I've tried this multiple times - but the result keeps the same and
shows me lots of troubles (the cluster is empty, no client has
accessed it)

#ceph -s
    cluster b04fc583-9e71-48b7-a741-92f4dff4cfef
     health HEALTH_WARN 470 pgs stale; 470 pgs stuck stale; 18 pgs
stuck unclean; 26 requests are blocked > 32 sec
     monmap e2: 3 mons at
{ceph-m-01=10.0.0.100:6789/0,ceph-m-02=10.0.1.101:6789/0,ceph-m-03=10.0.1.102:6789/0},
election epoch 8, quorum 0,1,2 ceph-m-01,ceph-m-02,ceph-m-03
     osdmap e409: 9 osds: 9 up, 9 in
      pgmap v1231: 480 pgs, 9 pools, 822 bytes data, 43 objects
            9373 MB used, 78317 GB / 78326 GB avail
                 451 stale+active+clean
                   1 stale+active+clean+scrubbing
                  10 active+clean
                  18 stale+active+remapped

_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to