I'd tried that sometime back but ran into some merge conflicts and was not sure who to turn to :) May I come to you for help with that?!
On Fri, Jun 17, 2016 at 3:29 PM, Atin Mukherjee <[email protected]> wrote: > > > On 06/17/2016 03:21 PM, B.K.Raghuram wrote: > > Thanks a ton Atin. That fixed cherry-pick. Will build it and let you > > know how it goes. Does it make sense to try and merge the whole upstream > > glusterfs repo for the 3.6 branch in order to get all the other bug > > fixes? That may bring in many more merge conflicts though.. > > Yup, I'd not recommend that. Applying your local changes on the latest > version is a much easier option :) > > > > > On Fri, Jun 17, 2016 at 3:07 PM, Atin Mukherjee <[email protected] > > <mailto:[email protected]>> wrote: > > > > I've resolved the merge conflicts and files are attached. Copy these > > files and follow the instructions from the cherry pick command which > > failed. > > > > ~Atin > > > > On 06/17/2016 02:55 PM, B.K.Raghuram wrote: > > > > > > Thanks Atin, I had three merge conflicts in the third patch.. I've > > > attached the files with the conflicts. Would any of the intervening > > > commits be needed as well? > > > > > > The conflicts were in : > > > > > > both modified: libglusterfs/src/mem-types.h > > > both modified: xlators/mgmt/glusterd/src/glusterd-utils.c > > > both modified: xlators/mgmt/glusterd/src/glusterd-utils.h > > > > > > > > > On Fri, Jun 17, 2016 at 2:17 PM, Atin Mukherjee < > [email protected] <mailto:[email protected]> > > > <mailto:[email protected] <mailto:[email protected]>>> wrote: > > > > > > > > > > > > On 06/17/2016 12:44 PM, B.K.Raghuram wrote: > > > > Thanks Atin.. I'm not familiar with pulling patches the > review system > > > > but will try:) > > > > > > It's not that difficult. Open the gerrit review link, go to > the download > > > drop box at the top right corner, click on it and then you > will see a > > > cherry pick option, copy that content and paste it the source > code repo > > > you host. If there are no merge conflicts, it should auto > apply, > > > otherwise you'd need to fix them manually. > > > > > > HTH. > > > Atin > > > > > > > > > > > On Fri, Jun 17, 2016 at 12:35 PM, Atin Mukherjee < > [email protected] <mailto:[email protected]> > > <mailto:[email protected] <mailto:[email protected]>> > > > > <mailto:[email protected] <mailto:[email protected]> > > <mailto:[email protected] <mailto:[email protected]>>>> wrote: > > > > > > > > > > > > > > > > On 06/16/2016 06:17 PM, Atin Mukherjee wrote: > > > > > > > > > > > > > > > On 06/16/2016 01:32 PM, B.K.Raghuram wrote: > > > > >> Thanks a lot Atin, > > > > >> > > > > >> The problem is that we are using a forked version of > 3.6.1 which has > > > > >> been modified to work with ZFS (for snapshots) but we > do not have the > > > > >> resources to port that over to the later versions of > gluster. > > > > >> > > > > >> Would you know of anyone who would be willing to take > this on?! > > > > > > > > > > If you can cherry pick the patches and apply them on > your source and > > > > > rebuild it, I can point the patches to you, but you'd > need to give a > > > > > day's time to me as I have some other items to finish > from my plate. > > > > > > > > > > > > Here is the list of the patches need to be applied on > the following > > > > order: > > > > > > > > http://review.gluster.org/9328 > > > > http://review.gluster.org/9393 > > > > http://review.gluster.org/10023 > > > > > > > > > > > > > > ~Atin > > > > >> > > > > >> Regards, > > > > >> -Ram > > > > >> > > > > >> On Thu, Jun 16, 2016 at 11:02 AM, Atin Mukherjee > > > > <[email protected] <mailto:[email protected]> > > <mailto:[email protected] <mailto:[email protected]>> > > > <mailto:[email protected] <mailto:[email protected]> > > <mailto:[email protected] <mailto:[email protected]>>> > > > > >> <mailto:[email protected] > > <mailto:[email protected]> <mailto:[email protected] > > <mailto:[email protected]>> > > > <mailto:[email protected] <mailto:[email protected]> > > <mailto:[email protected] <mailto:[email protected]>>>>> wrote: > > > > >> > > > > >> > > > > >> > > > > >> On 06/16/2016 10:49 AM, B.K.Raghuram wrote: > > > > >> > > > > > >> > > > > > >> > On Wed, Jun 15, 2016 at 5:01 PM, Atin Mukherjee > > > > <[email protected] <mailto:[email protected]> > > <mailto:[email protected] <mailto:[email protected]>> > > > <mailto:[email protected] <mailto:[email protected]> > > <mailto:[email protected] <mailto:[email protected]>>> > > > > <mailto:[email protected] <mailto:[email protected]> > > <mailto:[email protected] <mailto:[email protected]>> > > > <mailto:[email protected] <mailto:[email protected]> > > <mailto:[email protected] <mailto:[email protected]>>>> > > > > >> > <mailto:[email protected] > > <mailto:[email protected]> > > > <mailto:[email protected] <mailto:[email protected]>> > > <mailto:[email protected] <mailto:[email protected]> > > > <mailto:[email protected] <mailto:[email protected]>>> > > > > <mailto:[email protected] <mailto:[email protected]> > > <mailto:[email protected] <mailto:[email protected]>> > > > <mailto:[email protected] <mailto:[email protected]> > > <mailto:[email protected] <mailto:[email protected]>>>>>> wrote: > > > > >> > > > > > >> > > > > > >> > > > > > >> > On 06/15/2016 04:24 PM, B.K.Raghuram wrote: > > > > >> > > Hi, > > > > >> > > > > > > >> > > We're using gluster 3.6.1 and we > > periodically find > > > > that gluster commands > > > > >> > > fail saying the it could not get the lock > > on one of > > > > the brick machines. > > > > >> > > The logs on that machine then say > > something like : > > > > >> > > > > > > >> > > [2016-06-15 08:17:03.076119] E > > > > >> > > > [glusterd-op-sm.c:3058:glusterd_op_ac_lock] > > > > 0-management: Unable to > > > > >> > > acquire lock for vol2 > > > > >> > > > > > >> > This is a possible case if concurrent volume > > > operations > > > > are run. Do you > > > > >> > have any script which checks for volume > > status on an > > > > interval from all > > > > >> > the nodes, if so then this is an expected > > behavior. > > > > >> > > > > > >> > > > > > >> > Yes, I do have a couple of scripts that check on > > > volume and > > > > quota > > > > >> > status.. Given this, I do get a "Another > > transaction > > > is in > > > > progress.." > > > > >> > message which is ok. The problem is that > > sometimes I get > > > > the volume lock > > > > >> > held message which never goes away. This > sometimes > > > results > > > > in glusterd > > > > >> > consuming a lot of memory and CPU and the > > problem can > > > only > > > > be fixed with > > > > >> > a reboot. The log files are huge so I'm not > sure if > > > its ok > > > > to attach > > > > >> > them to an email. > > > > >> > > > > >> Ok, so this is known. We have fixed lots of stale > > lock > > > issues > > > > in 3.7 > > > > >> branch and some of them if not all were also > > backported to > > > > 3.6 branch. > > > > >> The issue is you are using 3.6.1 which is quite > > old. If you > > > > can upgrade > > > > >> to latest versions of 3.7 or at worst of 3.6 I am > > confident > > > > that this > > > > >> will go away. > > > > >> > > > > >> ~Atin > > > > >> > > > > > >> > > > > > > >> > > After sometime, glusterd then seems to > > give up > > > and die.. > > > > >> > > > > > >> > Do you mean glusterd shuts down or > > segfaults, if so I > > > > am more > > > > >> interested > > > > >> > in analyzing this part. Could you provide > > us the > > > > glusterd log, > > > > >> > cmd_history log file along with core (in > > case of > > > SEGV) from > > > > >> all the > > > > >> > nodes for the further analysis? > > > > >> > > > > > >> > > > > > >> > There is no segfault. glusterd just shuts down. > > As I said > > > > above, > > > > >> > sometimes this happens and sometimes it just > > continues to > > > > hog a lot of > > > > >> > memory and CPU.. > > > > >> > > > > > >> > > > > > >> > > > > > > >> > > Interestingly, I also find the following > line > > > in the > > > > >> beginning of > > > > >> > > etc-glusterfs-glusterd.vol.log and I dont > > know if > > > > this has any > > > > >> > > significance to the issue : > > > > >> > > > > > > >> > > [2016-06-14 06:48:57.282290] I > > > > >> > > > > [glusterd-store.c:2063:glusterd_restore_op_version] > > > > >> 0-management: > > > > >> > > Detected new install. Setting op-version > to > > > maximum : > > > > 30600 > > > > >> > > > > > > >> > > > > > >> > > > > > >> > What does this line signify? > > > > >> > > > > >> > > > > > > > > > > > > > > > > > > >
_______________________________________________ Gluster-users mailing list [email protected] http://www.gluster.org/mailman/listinfo/gluster-users
