On Fri, Jun 30, 2017 at 8:12 AM Nick Fisk <n...@fisk.me.uk> wrote:

> *From:* Alex Gorbachev [mailto:a...@iss-integration.com]
> *Sent:* 30 June 2017 03:54
> *To:* Ceph Users <ceph-users@lists.ceph.com>; n...@fisk.me.uk
>
>
> *Subject:* Re: [ceph-users] Kernel mounted RBD's hanging
>
>
>
>
>
> On Thu, Jun 29, 2017 at 10:30 AM Nick Fisk <n...@fisk.me.uk> wrote:
>
> Hi All,
>
> Putting out a call for help to see if anyone can shed some light on this.
>
> Configuration:
> Ceph cluster presenting RBD's->XFS->NFS->ESXi
> Running 10.2.7 on the OSD's and 4.11 kernel on the NFS gateways in a
> pacemaker cluster
> Both OSD's and clients are go into a pair of switches, single L2 domain (no
> sign from pacemaker that there is network connectivity issues)
>
> Symptoms:
> - All RBD's on a single client randomly hang for 30s to several minutes,
> confirmed by pacemaker and ESXi hosts complaining
> - Cluster load is minimal when this happens most times
> - All other clients with RBD's are not affected (Same RADOS pool), so its
> seems more of a client issue than cluster issue
> - It looks like pacemaker tries to also stop RBD+FS resource, but this also
> hangs
> - Eventually pacemaker succeeds in stopping resources and immediately
> restarts them, IO returns to normal
> - No errors, slow requests, or any other non normal Ceph status is reported
> on the cluster or ceph.log
> - Client logs show nothing apart from pacemaker
>
> Things I've tried:
> - Different kernels (potentially happened less with older kernels, but
> can't
> be 100% sure)
> - Disabling scrubbing and anything else that could be causing high load
> - Enabling Kernel RBD debugging (Problem maybe happens a couple of times a
> day, debug logging was not practical as I can't reproduce on demand)
>
> Anyone have any ideas?
>
>
>
> Nick, are you using any network aggregation, LACP?  Can you drop to a
> simplest possible configuration to make sure there's nothing on the network
> switch side?
>
>
>
> Hi Alex,
>
>
>
> The OSD nodes are active/backup bond and the active Nic on each one, all
> goes into the same switch. The NFS gateways are currently VM’s, but again
> the hypervisor is using the Nic on the same switch. The cluster and public
> networks are vlans on the same Nic and I don’t get any alerts from
> monitoring/pacemaker to suggest there are comms issues. But I will look
> into getting some ping logs done to see if they reveal anything.
>

Any chance this could be a hypervisor or VM-related issue?

Any possibility to run one gateway temporarily as a physical machine?


>
>
>
> Do you check the ceph.log for any anomalies?
>
>
>
> Yep, completely clean
>
>
>
> Any occurrences on OSD nodes, anything in their OSD logs or syslogs?
>
>
>
> Not that I can see. I’m using cache tiering, so all IO travels through a
> few OSD’s. I guess this might make it easier to try and see whats going on.
> But the random nature of it, means it’s not always easy to catch.
>
>
>
> Aany odd page cache settings on the clients?
>
>
>
> The only customizations on the clients are readahead, some TCP tunings and
> min free kbytes.
>
>
>
> Alex
>
>
>
>
>
> Thanks,
> Nick
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> <http://xo4t.mj.am/lnk/AEUAMGSsyuUAAAAAAAAAAFhNkjYAADNJBWwAAAAAAACRXwBZVkBBEimV6rRsR9ueEOKOWc4YEwAAlBI/1/KaykvSTe4bVbKn7nnq-msA/aHR0cDovL2xpc3RzLmNlcGguY29tL2xpc3RpbmZvLmNnaS9jZXBoLXVzZXJzLWNlcGguY29t>
>
> --
>
> --
>
> Alex Gorbachev
>
> Storcium
>
> --
--
Alex Gorbachev
Storcium
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to