On Thursday 22 March 2018 01:07 PM, Atin Mukherjee wrote:
On Thu, Mar 22, 2018 at 12:38 PM, Jiffin Tony Thottan <[email protected] <mailto:[email protected]>> wrote:On Thursday 22 March 2018 12:29 PM, Jiffin Tony Thottan wrote:On Wednesday 21 March 2018 09:06 AM, Atin Mukherjee wrote:On Wed, Mar 21, 2018 at 12:18 AM, Shyam Ranganathan <[email protected] <mailto:[email protected]>> wrote: On 03/20/2018 01:10 PM, Jiffin Thottan wrote: > Hi Shyam, > > Actually I planned to do the release on March 8th(posted the release note on that day). But it didn't happen. > I didn't merge any patches post sending the release note(blocker bug had some merge conflict issue on that so I skipped AFAIR). > I performed 3.12.7 tagging yesterday and ran the build job today. > > Can u please provide a suggestion here ? Do I need to perform a 3.12.7-1 for the blocker bug ? I see that the bug is marked against the tracker, but is not a regression or an issue that is serious enough that it cannot wait for the next minor release. Copied Atin to the mail, who opened that issue for his comments. If he agrees, let's get this moving and get the fix into the next minor release. Even though it's not a regression and a day 1 bug with brick multiplexing, the issue is severe enough to consider this to be fixed *asap* . In this scenario, if you're running a multi node cluster with brick multiplexing enabled and one node down and there're some volume operations performed and post that when the node comes back, brick processes fail to come up.Issue is impact only with glusterd, whether any other component needs this fix?Sorry I meant brick multiplexing not glusterd -- JiffinIf it is issue not report from upstream user/community, I prefer to take it for next release.IMO, assessment of an issue should be done based on its merit, not based on where it originates from. It might be a fair question to ask that "do we have users who have brick multiplexing enabled" and based on that take a call to fix it immediately or as part of next update but at the same time, you're still exposing a known problem with out flagging a warning that don't use brick multiplexing till this bug is fixed.
I have not yet sent the announcement mail for the release nor sent release notes to https://docs.gluster.org/en. I can mention about it over there
-- Jiffin
Regards, Jiffin> > -- > Regards, > Jiffin > > > > > ----- Original Message ----- > From: "Shyam Ranganathan" <[email protected] <mailto:[email protected]>> > To: [email protected] <mailto:[email protected]>, [email protected] <mailto:[email protected]>, [email protected] <mailto:[email protected]> > Sent: Tuesday, March 20, 2018 9:06:57 PM > Subject: Re: [Gluster-Maintainers] glusterfs-3.12.7 released > > On 03/20/2018 11:19 AM, [email protected] <mailto:[email protected]> wrote: >> SRC: https://build.gluster.org/job/release-new/47/artifact/glusterfs-3.12.7.tar.gz <https://build.gluster.org/job/release-new/47/artifact/glusterfs-3.12.7.tar.gz> >> HASH: https://build.gluster.org/job/release-new/47/artifact/glusterfs-3.12.7.sha512sum <https://build.gluster.org/job/release-new/47/artifact/glusterfs-3.12.7.sha512sum> >> >> This release is made off jenkins-release-47 > > Jiffin, there are about 6 patches ready in the 3.12 queue, that are not > merged for this release, why? > https://review.gluster.org/#/projects/glusterfs,dashboards/dashboard:3-12-dashboard <https://review.gluster.org/#/projects/glusterfs,dashboards/dashboard:3-12-dashboard> > > The tracker bug for 3.12.7 calls out > https://bugzilla.redhat.com/show_bug.cgi?id=1543708 <https://bugzilla.redhat.com/show_bug.cgi?id=1543708> as a blocker, and > has a patch, which is not merged. > > Was this some test packaging job? > > > > >> >> >> >> _______________________________________________ >> maintainers mailing list >> [email protected] <mailto:[email protected]> >> http://lists.gluster.org/mailman/listinfo/maintainers <http://lists.gluster.org/mailman/listinfo/maintainers> >> > _______________________________________________ > maintainers mailing list > [email protected] <mailto:[email protected]> > http://lists.gluster.org/mailman/listinfo/maintainers <http://lists.gluster.org/mailman/listinfo/maintainers> >_______________________________________________ maintainers mailing list [email protected] <mailto:[email protected]> http://lists.gluster.org/mailman/listinfo/maintainers <http://lists.gluster.org/mailman/listinfo/maintainers>_______________________________________________ maintainers mailing list [email protected] <mailto:[email protected]> http://lists.gluster.org/mailman/listinfo/maintainers <http://lists.gluster.org/mailman/listinfo/maintainers>
_______________________________________________ maintainers mailing list [email protected] http://lists.gluster.org/mailman/listinfo/maintainers
