Hello,

On Mon, 15 Jun 2015 23:11:07 +0000 Matteo Dacrema wrote:

> With 3.16.3 kernel it seems to be stable but I've discovered one new
> issue.
> 
> If I take down one of the two osd node all the client stop to respond.
> 
How did you take the node down?

What is your "osd_pool_default_min_size"?

Penultimately, you wouldn't deploy a cluster with just 2 storage nodes in
production anyway.

Christian
> 
> Here the output of ceph -s
> 
> ceph -s
>     cluster 2de7b17f-0a3e-4109-b878-c035dd2f7735
>      health HEALTH_WARN
>             256 pgs degraded
>             127 pgs stuck inactive
>             127 pgs stuck unclean
>             256 pgs undersized
>             recovery 1457662/2915324 objects degraded (50.000%)
>             4/8 in osds are down
>             clock skew detected on mon.cephmds01, mon.ceph-mon1
>      monmap e5: 3 mons at
> {ceph-mon1=10.29.81.184:6789/0,cephmds01=10.29.81.161:6789/0,cephmds02=10.29.81.160:6789/0}
> election epoch 64, quorum 0,1,2 cephmds02,cephmds01,ceph-mon1 mdsmap
> e176: 1/1/1 up {0=cephmds01=up:active}, 1 up:standby osdmap e712: 8
> osds: 4 up, 8 in pgmap v420651: 256 pgs, 2 pools, 133 GB data, 1423
> kobjects 289 GB used, 341 GB / 631 GB avail
>             1457662/2915324 objects degraded (50.000%)
>                  256 undersized+degraded+peered
>   client io 86991 B/s wr, 0 op/s
> 
> 
> When I take UP the node all clients resume to work.
> 
> Thanks,
> Matteo
> 
> ?
> 
> 
> ________________________________
> Da: ceph-users <[email protected]> per conto di Matteo
> Dacrema <[email protected]> Inviato: luned? 15 giugno 2015 12:37
> A: John Spray; Lincoln Bryant; ceph-users
> Oggetto: Re: [ceph-users] CephFS client issue
> 
> 
> Ok, I'll update kernel to 3.16.3 version and let you know.
> 
> 
> Thanks,
> 
> Matteo
> 
> ________________________________
> Da: John Spray <[email protected]>
> Inviato: luned? 15 giugno 2015 10:51
> A: Matteo Dacrema; Lincoln Bryant; ceph-users
> Oggetto: Re: [ceph-users] CephFS client issue
> 
> 
> 
> On 14/06/15 20:00, Matteo Dacrema wrote:
> 
> Hi Lincoln,
> 
> 
> I'm using the kernel client.
> 
> Kernel version is: 3.13.0-53-generic?
> 
> That's old by CephFS standards.  It's likely that the issue you're
> seeing is one of the known bugs (which were actually the motivation for
> adding the warning message you're seeing).
> 
> John
> 
> --
> Questo messaggio e' stato analizzato con Libra ESVA ed e' risultato non
> infetto. Clicca qui per segnalarlo come
> spam.<http://esva01.enter.it/cgi-bin/learn-msg.cgi?id=637D840263.A210F>
> Clicca qui per metterlo in
> blacklist<http://esva01.enter.it/cgi-bin/learn-msg.cgi?blacklist=1&id=637D840263.A210F>
> 
> --
> Questo messaggio e' stato analizzato con Libra ESVA ed e' risultato non
> infetto. Clicca qui per segnalarlo come
> spam.<http://esva01.enter.it/cgi-bin/learn-msg.cgi?id=DBA3140262.AE1FB>
> Clicca qui per metterlo in
> blacklist<http://esva01.enter.it/cgi-bin/learn-msg.cgi?blacklist=1&id=DBA3140262.AE1FB>


-- 
Christian Balzer        Network/Systems Engineer                
[email protected]           Global OnLine Japan/Fusion Communications
http://www.gol.com/
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to