Re: [Gluster-users] Volume not healing

2021-03-22 Thread Diego Zuccato
Il 19/03/21 18:06, Strahil Nikolov ha scritto: > Are you running it against the fuse mountpoint ? Yup. > You are not supposed to see 'no such file or directory' ... Maybe > something more serious is going on. Between that and the duplicated files,that's for sure. But I don't know where to look

Re: [Gluster-users] Volume not healing

2021-03-22 Thread Diego Zuccato
Il 20/03/21 15:21, Zenon Panoussis ha scritto: > When you have 0 files that need healing, > gluster volume heal BigVol granular-entry-heal enable > I have tested with and without granular and, empirically, > without any hard statistics, I find granular considerably > faster. Tks for the hint,

Re: [Gluster-users] Volume not healing

2021-03-20 Thread Zenon Panoussis
> Is it possible to speed it up? Nodes are nearly idle... When you have 0 files that need healing, gluster volume heal BigVol granular-entry-heal enable I have tested with and without granular and, empirically, without any hard statistics, I find granular considerably faster.

Re: [Gluster-users] Volume not healing

2021-03-19 Thread Diego Zuccato
Il 19/03/21 13:17, Strahil Nikolov ha scritto: > find /FUSE/mountpoint -exec stat {} \; Running it now (redirecting stdout to /dev/null). It's finding quite a lot of "no such file or directory" errors. -- Diego Zuccato DIFA - Dip. di Fisica e Astronomia Servizi Informatici Alma Mater Studiorum

Re: [Gluster-users] Volume not healing

2021-03-19 Thread Diego Zuccato
Il 19/03/21 11:06, Diego Zuccato ha scritto: > I tried to run "gluster v heal BigVol info summary" and got quite a high > count of entries to be healed on some bricks: > # gluster v heal BigVol info summary|grep pending|grep -v ' 0$' > Number of entries in heal pending: 41 > Number of entries in

[Gluster-users] Volume not healing

2021-03-19 Thread Diego Zuccato
Hello all. I have a "problematic" volume. It was Rep3a1 with a dedicated VM for the arbiters. Too bad I understimated RAM needs and the arbiters VM crashed frequently for OOM (had just 8GB allocated). Even the other two nodes sometimes crashed, too, during a remove-brick operation (other