So that's two hosts - if this is a new cluster chances are the pools
have replication size=3, and won't place replica pgs on the same host...
'ceph osd dump' will let you know if this is the case. If it is ether
reduce size to 2, add another host or edit your crush rules to allow
replica pgs on the same host.
Cheers
Mark
On 09/05/14 18:20, Georg Höllrigl wrote:
#ceph osd tree
# id weight type name up/down reweight
-1 76.47 root default
-2 32.72 host ceph-s-01
0 7.27 osd.0 up 1
1 7.27 osd.1 up 1
2 9.09 osd.2 up 1
3 9.09 osd.3 up 1
-3 43.75 host ceph-s-02
4 10.91 osd.4 up 1
5 0.11 osd.5 up 1
6 10.91 osd.6 up 1
7 10.91 osd.7 up 1
8 10.91 osd.8 up 1
On 08.05.2014 19:11, Craig Lewis wrote:
What does `ceph osd tree` output?
On 5/8/14 07:30 , Georg Höllrigl wrote:
Hello,
We've a fresh cluster setup - with Ubuntu 14.04 and ceph firefly. By
now I've tried this multiple times - but the result keeps the same and
shows me lots of troubles (the cluster is empty, no client has
accessed it)
#ceph -s
cluster b04fc583-9e71-48b7-a741-92f4dff4cfef
health HEALTH_WARN 470 pgs stale; 470 pgs stuck stale; 18 pgs
stuck unclean; 26 requests are blocked > 32 sec
monmap e2: 3 mons at
{ceph-m-01=10.0.0.100:6789/0,ceph-m-02=10.0.1.101:6789/0,ceph-m-03=10.0.1.102:6789/0},
election epoch 8, quorum 0,1,2 ceph-m-01,ceph-m-02,ceph-m-03
osdmap e409: 9 osds: 9 up, 9 in
pgmap v1231: 480 pgs, 9 pools, 822 bytes data, 43 objects
9373 MB used, 78317 GB / 78326 GB avail
451 stale+active+clean
1 stale+active+clean+scrubbing
10 active+clean
18 stale+active+remapped
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com