On Thu, Mar 26, 2015 at 4:40 PM, Gregory Farnum <g...@gregs42.com> wrote:

> Has the OSD actually been detected as down yet?
>
>
I believe it has, however I can't directly check because "ceph health"
starts to hang when I down the second node.


> You'll also need to set that min size on your existing pools ("ceph
> osd pool <pool> set min_size 1" or similar) to change their behavior;
> the config option only takes effect for newly-created pools. (Thus the
> "default".)


I've done this, however the behavior is the same:

$ for f in `ceph osd lspools | sed 's/[0-9]//g' | sed 's/,//g'`; do ceph
osd pool set $f min_size 1; done
set pool 0 min_size to 1
set pool 1 min_size to 1
set pool 2 min_size to 1
set pool 3 min_size to 1
set pool 4 min_size to 1
set pool 5 min_size to 1
set pool 6 min_size to 1
set pool 7 min_size to 1

$ ceph -w
    cluster db460aa2-5129-4aaa-8b2e-43eac727124e
     health HEALTH_WARN 1 mons down, quorum 0,1 ceph-node-1,ceph-node-2
     monmap e3: 3 mons at {ceph-node-1=
192.168.122.121:6789/0,ceph-node-2=192.168.122.131:6789/0,ceph-node-3=192.168.122.141:6789/0},
election epoch 194, quorum 0,1 ceph-node-1,ceph-node-2
     mdsmap e94: 1/1/1 up {0=ceph-node-1=up:active}
     osdmap e362: 3 osds: 2 up, 2 in
      pgmap v5913: 840 pgs, 8 pools, 7441 MB data, 994 objects
            25329 MB used, 12649 MB / 40059 MB avail
                 840 active+clean

2015-03-26 17:23:56.009938 mon.0 [INF] pgmap v5913: 840 pgs: 840
active+clean; 7441 MB data, 25329 MB used, 12649 MB / 40059 MB avail
2015-03-26 17:25:51.042802 mon.0 [INF] pgmap v5914: 840 pgs: 840
active+clean; 7441 MB data, 25329 MB used, 12649 MB / 40059 MB avail; 0 B/s
rd, 260 kB/s wr, 13 op/s
2015-03-26 17:25:56.046491 mon.0 [INF] pgmap v5915: 840 pgs: 840
active+clean; 7441 MB data, 25333 MB used, 12645 MB / 40059 MB avail; 0 B/s
rd, 943 kB/s wr, 38 op/s
2015-03-26 17:26:01.058167 mon.0 [INF] pgmap v5916: 840 pgs: 840
active+clean; 7441 MB data, 25335 MB used, 12643 MB / 40059 MB avail; 0 B/s
rd, 10699 kB/s wr, 621 op/s

<this is where i kill the second OSD>

2015-03-26 17:26:26.778461 7f4ebeffd700  0 monclient: hunting for new mon
2015-03-26 17:26:30.701099 7f4ec45f5700  0 -- 192.168.122.111:0/1007741 >>
192.168.122.141:6789/0 pipe(0x7f4ec0023200 sd=3 :0 s=1 pgs=0 cs=0 l=1
c=0x7f4ec0023490).fault
2015-03-26 17:26:42.701154 7f4ec44f4700  0 -- 192.168.122.111:0/1007741 >>
192.168.122.131:6789/0 pipe(0x7f4ec00251b0 sd=3 :0 s=1 pgs=0 cs=0 l=1
c=0x7f4ec0025440).fault

And all writes block until I bring back an OSD.

Lee
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to