Executing "gluster volume set all cluster.op-version <op-version>"on all the existing nodes will solve this problem.
If issue still persists please provide me following logs (working-cluster + newly added peer) 1. glusterd.info file from /var/lib/glusterd from all nodes 2. glusterd.logs from all nodes 3. info file from all the nodes. 4. cmd-history from all the nodes. Thanks Gaurav On Mon, Sep 4, 2017 at 2:09 PM, lejeczek <[email protected]> wrote: > I do not see, did you write anything? > > On 03/09/17 11:54, Gaurav Yadav wrote: > >> >> >> On Fri, Sep 1, 2017 at 9:02 PM, lejeczek <[email protected] <mailto: >> [email protected]>> wrote: >> >> you missed my reply before? >> here: >> >> now, a "weir" thing >> >> I did that, still fourth peer rejected, still fourth >> probe would fail to restart(all after upping to 31004) >> I redone, wiped and re-probed from a different peer >> than I did from first time and now not rejected. >> now I'm restarting fourth(newly added) peer's glusterd >> and.. it seems to work.(even though tier-enabled=0 is >> still there, now on all four peers, was not there on >> three before working peers) >> >> I there anything I should double check & make sure all >> is 100% fine before I use that newly added peer for >> bricks? >> >> For this only I need logs to see what has gone wrong. >> >> >> Please provide me following things (working-cluster + newly added peer) >> 1. glusterd.info <http://glusterd.info> <http://glusterd.info> file from >> /var/lib/glusterd from all nodes >> 2. glusterd.logs from all nodes >> 3. info file from all the nodes. >> 4. cmd-history from all the nodes. >> >> >> On 01/09/17 11:11, Gaurav Yadav wrote: >> >> I replicate the problem locally and with the steps >> I suggested you, it worked for me... >> >> Please provide me following things >> (working-cluster + newly added peer) >> 1. glusterd.info <http://glusterd.info> >> <http://glusterd.info> file from /var/lib/glusterd >> from all nodes >> 2. glusterd.logs from all nodes >> 3. info file from all the nodes. >> 4. cmd-history from all the nodes. >> >> >> On Fri, Sep 1, 2017 at 3:39 PM, lejeczek >> <[email protected] <mailto:[email protected]> >> <mailto:[email protected] >> <mailto:[email protected]>>> wrote: >> >> Like I said, I upgraded from 3.8 to 3.10 a >> while ago, >> at the moment 3.10.5, only now with 3.10.5 I >> tried to >> add a peer. >> >> On 01/09/17 10:51, Gaurav Yadav wrote: >> >> What is gluster --version on all these nodes? >> >> On Fri, Sep 1, 2017 at 3:18 PM, lejeczek >> <[email protected] >> <mailto:[email protected]> >> <mailto:[email protected] >> <mailto:[email protected]>> >> <mailto:[email protected] >> <mailto:[email protected]> >> >> <mailto:[email protected] >> <mailto:[email protected]>>>> wrote: >> >> on first node I got >> $ gluster volume set all >> cluster.op-version 31004 >> volume set: failed: Commit failed on >> 10.5.6.49. Please >> check log file for details. >> >> but I immediately proceeded to >> remaining nodes >> and: >> >> $ gluster volume get all >> cluster.op-version >> Option Value >> ------ ----- >> cluster.op-version 30712 >> $ gluster volume set all >> cluster.op-version 31004 >> volume set: failed: Required op-version >> (31004) should >> not be equal or lower than current cluster >> op-version >> (31004). >> $ gluster volume get all >> cluster.op-version >> Option Value >> ------ ----- >> cluster.op-version 31004 >> >> last, third node: >> >> $ gluster volume get all >> cluster.op-version >> Option Value >> ------ ----- >> cluster.op-version 30712 >> $ gluster volume set all >> cluster.op-version 31004 >> volume set: failed: Required op-version >> (31004) should >> not be equal or lower than current cluster >> op-version >> (31004). >> $ gluster volume get all >> cluster.op-version >> Option Value >> ------ ----- >> cluster.op-version 31004 >> >> So, even though it failed as above, >> now I see that >> it's 31004 on all three peers, at least >> according to >> "volume get all cluster.op-version" >> command. >> >> >> On 01/09/17 10:38, Gaurav Yadav wrote: >> >> gluster volume set all >> cluster.op-version >> 31004 >> >> >> >> >> >> >> >> >
_______________________________________________ Gluster-users mailing list [email protected] http://lists.gluster.org/mailman/listinfo/gluster-users
