On Thursday 22 March 2018 01:38 PM, Jiffin Tony Thottan wrote:
On Thursday 22 March 2018 01:07 PM, Atin Mukherjee wrote:
On Thu, Mar 22, 2018 at 12:38 PM, Jiffin Tony Thottan
<jthot...@redhat.com <mailto:jthot...@redhat.com>> wrote:
On Thursday 22 March 2018 12:29 PM, Jiffin Tony Thottan wrote:
On Wednesday 21 March 2018 09:06 AM, Atin Mukherjee wrote:
On Wed, Mar 21, 2018 at 12:18 AM, Shyam Ranganathan
<srang...@redhat.com <mailto:srang...@redhat.com>> wrote:
On 03/20/2018 01:10 PM, Jiffin Thottan wrote:
> Hi Shyam,
>
> Actually I planned to do the release on March 8th(posted
the release note on that day). But it didn't happen.
> I didn't merge any patches post sending the release
note(blocker bug had some merge conflict issue on that so I
skipped AFAIR).
> I performed 3.12.7 tagging yesterday and ran the build
job today.
>
> Can u please provide a suggestion here ? Do I need to
perform a 3.12.7-1 for the blocker bug ?
I see that the bug is marked against the tracker, but is not a
regression or an issue that is serious enough that it
cannot wait for
the next minor release.
Copied Atin to the mail, who opened that issue for his
comments. If he
agrees, let's get this moving and get the fix into the next
minor release.
Even though it's not a regression and a day 1 bug with brick
multiplexing, the issue is severe enough to consider this to be
fixed *asap* . In this scenario, if you're running a multi node
cluster with brick multiplexing enabled and one node down and
there're some volume operations performed and post that when
the node comes back, brick processes fail to come up.
Issue is impact only with glusterd, whether any other component
needs this fix?
Sorry I meant brick multiplexing not glusterd
--
Jiffin
If it is issue not report from upstream user/community, I prefer
to take it for next release.
IMO, assessment of an issue should be done based on its merit, not
based on where it originates from. It might be a fair question to ask
that "do we have users who have brick multiplexing enabled" and based
on that take a call to fix it immediately or as part of next update
but at the same time, you're still exposing a known problem with out
flagging a warning that don't use brick multiplexing till this bug is
fixed.
I have not yet sent the announcement mail for the release nor sent
release notes to https://docs.gluster.org/en. I can mention about it
over there
--
Jiffin
Can u please tell me whether it works for u?
--
Jiffin
Regards,
Jiffin
>
> --
> Regards,
> Jiffin
>
>
>
>
> ----- Original Message -----
> From: "Shyam Ranganathan" <srang...@redhat.com
<mailto:srang...@redhat.com>>
> To: jenk...@build.gluster.org
<mailto:jenk...@build.gluster.org>, packag...@gluster.org
<mailto:packag...@gluster.org>, maintainers@gluster.org
<mailto:maintainers@gluster.org>
> Sent: Tuesday, March 20, 2018 9:06:57 PM
> Subject: Re: [Gluster-Maintainers] glusterfs-3.12.7 released
>
> On 03/20/2018 11:19 AM, jenk...@build.gluster.org
<mailto:jenk...@build.gluster.org> wrote:
>> SRC:
https://build.gluster.org/job/release-new/47/artifact/glusterfs-3.12.7.tar.gz
<https://build.gluster.org/job/release-new/47/artifact/glusterfs-3.12.7.tar.gz>
>> HASH:
https://build.gluster.org/job/release-new/47/artifact/glusterfs-3.12.7.sha512sum
<https://build.gluster.org/job/release-new/47/artifact/glusterfs-3.12.7.sha512sum>
>>
>> This release is made off jenkins-release-47
>
> Jiffin, there are about 6 patches ready in the 3.12
queue, that are not
> merged for this release, why?
>
https://review.gluster.org/#/projects/glusterfs,dashboards/dashboard:3-12-dashboard
<https://review.gluster.org/#/projects/glusterfs,dashboards/dashboard:3-12-dashboard>
>
> The tracker bug for 3.12.7 calls out
> https://bugzilla.redhat.com/show_bug.cgi?id=1543708
<https://bugzilla.redhat.com/show_bug.cgi?id=1543708> as a
blocker, and
> has a patch, which is not merged.
>
> Was this some test packaging job?
>
>
>
>
>>
>>
>>
>> _______________________________________________
>> maintainers mailing list
>> maintainers@gluster.org <mailto:maintainers@gluster.org>
>> http://lists.gluster.org/mailman/listinfo/maintainers
<http://lists.gluster.org/mailman/listinfo/maintainers>
>>
> _______________________________________________
> maintainers mailing list
> maintainers@gluster.org <mailto:maintainers@gluster.org>
> http://lists.gluster.org/mailman/listinfo/maintainers
<http://lists.gluster.org/mailman/listinfo/maintainers>
>
_______________________________________________
maintainers mailing list
maintainers@gluster.org <mailto:maintainers@gluster.org>
http://lists.gluster.org/mailman/listinfo/maintainers
<http://lists.gluster.org/mailman/listinfo/maintainers>
_______________________________________________
maintainers mailing list
maintainers@gluster.org <mailto:maintainers@gluster.org>
http://lists.gluster.org/mailman/listinfo/maintainers
<http://lists.gluster.org/mailman/listinfo/maintainers>
_______________________________________________
maintainers mailing list
maintainers@gluster.org
http://lists.gluster.org/mailman/listinfo/maintainers
_______________________________________________
maintainers mailing list
maintainers@gluster.org
http://lists.gluster.org/mailman/listinfo/maintainers