Does the patch in KAFKA-1382 apply on the 0.8.1 branch?  If not if you
could make a patch that does would be great.

I will kick off a discussion for KAFKA-1382 and the scala 2.11 for 0.8.1.2
release (and see what others may think we should do like the gradle changes
I think we should do too for src release issues (and the jars in the
repo)).  I will send that on dev/user in a little bit (please comment +1
community support please on that thread for the release).

/*******************************************
 Joe Stein
 Founder, Principal Consultant
 Big Data Open Source Security LLC
 http://www.stealth.ly
 Twitter: @allthingshadoop <http://www.twitter.com/allthingshadoop>
********************************************/

On Tue, Sep 30, 2014 at 11:10 AM, Seshadri, Balaji <balaji.sesha...@dish.com
> wrote:

> I would love to help you guys to make Kafka best in Pub/Sub, will continue
> doing that whenever I can.
>
> Do we have 0.8.1.2 release tag  or should we apply patch on top of 0.8.1.1
> tag because we need this KAFKA-1382 JIRA ?.
>
> Balaji
>
> From: Neha Narkhede [mailto:neha.narkh...@gmail.com]
> Sent: Monday, September 29, 2014 5:21 PM
> To: Seshadri, Balaji
> Cc: users@kafka.apache.org
> Subject: Re: BadVersion state in Kafka Logs
>
> It is difficult to predict an exact date. Though all the discussions of
> the progress and ETA are on the mailing list. You can follow the
> discussions to know the details and/or offer to help out on the outstanding
> issues.
>
> On Mon, Sep 29, 2014 at 3:48 PM, Seshadri, Balaji <
> balaji.sesha...@dish.com<mailto:balaji.sesha...@dish.com>> wrote:
> Neha,
>
> Do you know the date in Oct when 0.8.2 is going to be out ?.
>
> Thanks,
>
> Balaji
>
> From: Neha Narkhede [mailto:neha.narkh...@gmail.com<mailto:
> neha.narkh...@gmail.com>]
> Sent: Thursday, September 25, 2014 1:08 PM
> To: Seshadri, Balaji
> Cc: users@kafka.apache.org<mailto:users@kafka.apache.org>
>
> Subject: Re: BadVersion state in Kafka Logs
>
> We are close to the release. I'd probably expect 0.8.2 sometime in October.
>
> On Thu, Sep 25, 2014 at 10:37 AM, Seshadri, Balaji <
> balaji.sesha...@dish.com<mailto:balaji.sesha...@dish.com>> wrote:
> Hi Neha,
>
> Do you know when are you guys releasing 0.8.2 ?.
>
> Thanks,
>
> Balaji
>
> -----Original Message-----
> From: Seshadri, Balaji [mailto:balaji.sesha...@dish.com<mailto:
> balaji.sesha...@dish.com>]
> Sent: Thursday, September 25, 2014 9:41 AM
> To: users@kafka.apache.org<mailto:users@kafka.apache.org>
> Subject: RE: BadVersion state in Kafka Logs
>
> Thanks for the replay.
>
> Please let me know if we can use trunk as 0.8.2 is not yet released.
>
> Balaji
> ________________________________________
> From: Neha Narkhede [neha.narkh...@gmail.com<mailto:
> neha.narkh...@gmail.com>]
> Sent: Wednesday, September 24, 2014 6:32 PM
> To: users@kafka.apache.org<mailto:users@kafka.apache.org>
> Subject: Re: BadVersion state in Kafka Logs
>
> From the logs you've attached, my guess is it's most likely due to
> KAFKA-1382.
>
> Thanks,
> Neha
>
> On Wed, Sep 24, 2014 at 10:48 AM, Seshadri, Balaji <
> balaji.sesha...@dish.com<mailto:balaji.sesha...@dish.com>
> > wrote:
>
> > Hi,
> >
> >
> >
> > We got the below error in our logs and our consumers stopped consuming
> > any data ?.It worked only after restart.
> >
> >
> >
> > We would like to confirm that it's because we are running with
> > 0.8-beta version and not 0.8 release version to convince "THE MGMT" guys.
> >
> >
> >
> > Please let me know if it's this KAFKA-1382 causing the issue.
> >
> >
> >
> > Thanks,
> >
> >
> >
> > Balaji
> >
> >
> >
> > *From:* Gulia, Vikram
> > *Sent:* Wednesday, September 24, 2014 8:43 AM
> > *To:* Sharma, Navdeep; #IT-MAD DES; #IT-MAA
> > *Cc:* Alam, Mohammad Shah
> > *Subject:* RE: 9/23 prod issue - offline kafka partitions.
> >
> >
> >
> > Adding full MAA distro.
> >
> >
> >
> > DES Offshore looked in to the logs on kafka servers and seems like the
> > issue we encountered yesterday may be described in these threads,
> > please have a look -
> >
> >
> >
> > http://permalink.gmane.org/gmane.comp.apache.kafka.user/1904
> >
> >
> >
> > https://issues.apache.org/jira/browse/KAFKA-1382 (it describes the
> > fix/patch which is available in 0.8.1.2/0.8.2<http://0.8.1.2/0.8.2>)
> >
> >
> >
> > Thank You,
> >
> > Vikram Gulia
> >
> >
> >
> > *From:* Sharma, Navdeep
> > *Sent:* Wednesday, September 24, 2014 6:53 AM
> > *To:* Gulia, Vikram; #IT-MAD DES
> > *Cc:* #IT-MAA Offshore; Alam, Mohammad Shah
> > *Subject:* RE: 9/23 prod issue - offline kafka partitions.
> >
> >
> >
> > Hi Vikram,
> >
> >
> >
> > We analyzed  below mentioned issue with MAA-Offshore (Abhishek) and
> > found that the error occurred only on 23 Sept. This is  not historical
> > as we checked last 4 days logs.
> >
> >
> >
> > It looks like that consumer got stopped on September 22 2014 for Linux
> > patching activty.MAA started consumer September  23 2014 at 1:00 AM.
> >
> >
> >
> > *Issue *in  server log   *"* *BadVersion for
> > /brokers/topics/rain-burn-in/partitions/121/state"*  but it is not
> > present in previous 4 days logs.
> >
> > More detail of this error can be found at-
> >
> > http://permalink.gmane.org/gmane.comp.apache.kafka.user/1904
> >
> >
> >
> > We are not sure about data loss in this scenario and working on this.
> >
> >
> >
> >
> >
> >
> >
> > Let us know if any concerns.
> >
> >
> >
> > [image: cid:image001.gif@01CF7B0A.03F21580]
> >
> > Navdeep Sharma
> > Developer - offshore,  Middleware Applications & Development o India:
> > 0120-4532000 - 2234
> > c: +91-9911698102<tel:%2B91-9911698102>
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> >
> > *From:* Gulia, Vikram
> > *Sent:* Tuesday, September 23, 2014 6:17 PM
> > *To:* #IT-MAD DES
> > *Subject:* FW: 9/23 prod issue - offline kafka partitions.
> >
> >
> >
> > DES Offshore dev,
> >
> >
> >
> > Please work with MAA offshore to monitor the kafka broker as we had
> > this incident where lot of partitions went offline around 1.45 PM MST
> > and MAA has to restart the kafka servers. We may have lost messages
> > and we need to see if there is a way to figure out what was the impact.
> >
> >
> >
> > Also, check the logs for kafka servers and see if we can figure out
> > why did partitions go offline or are un-available? Let us know if you
> > find anything relevant.
> >
> >
> >
> > Thank You,
> >
> > Vikram Gulia
> >
> >
> >
> > *From:* Nielsen, Andy
> > *Sent:* Tuesday, September 23, 2014 5:04 PM
> > *To:* #IT-MAD DES; Gulia, Vikram
> > *Cc:* #IT-MAA
> > *Subject:* 9/23 prod issue - offline kafka partitions.
> >
> >
> >
> > desadmin@pc1mwdpl01 ~/bin $ ./kafka.sh topic --unavailable-partitions
> >
> > topic: account-access   partition: 21   leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: account-access   partition: 51   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: account-access   partition: 81   leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: account-access   partition: 111  leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: account-activated        partition: 13   leader: -1      replicas:
> > 4,6,1 isr:
> >
> > topic: account-activated        partition: 43   leader: -1      replicas:
> > 4,6,1 isr:
> >
> > topic: account-activated        partition: 73   leader: -1      replicas:
> > 4,6,1 isr:
> >
> > topic: account-activated        partition: 103  leader: -1      replicas:
> > 4,6,1 isr: 1
> >
> > topic: account-adjustment-issued        partition: 27   leader: -1
> > replicas: 4,6,1 isr:
> >
> > topic: account-adjustment-issued        partition: 57   leader: -1
> > replicas: 4,6,1 isr:
> >
> > topic: account-adjustment-issued        partition: 87   leader: -1
> > replicas: 4,6,1 isr: 1
> >
> > topic: account-adjustment-issued        partition: 117  leader: -1
> > replicas: 4,6,1 isr:
> >
> > topic: account-created  partition: 11   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: account-created  partition: 41   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: account-created  partition: 71   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: account-created  partition: 101  leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: account-info-updated     partition: 7    leader: -1      replicas:
> > 4,6,1 isr: 1
> >
> > topic: account-info-updated     partition: 37   leader: -1      replicas:
> > 4,6,1 isr: 1
> >
> > topic: account-info-updated     partition: 67   leader: -1      replicas:
> > 4,6,1 isr:
> >
> > topic: account-info-updated     partition: 97   leader: -1      replicas:
> > 4,6,1 isr: 1
> >
> > topic: account-info-updated     partition: 127  leader: -1      replicas:
> > 4,6,1 isr: 1
> >
> > topic: application-access       partition: 21   leader: -1      replicas:
> > 4,6,1 isr: 1
> >
> > topic: application-access       partition: 51   leader: -1      replicas:
> > 4,6,1 isr: 1
> >
> > topic: application-access       partition: 81   leader: -1      replicas:
> > 4,6,1 isr: 1
> >
> > topic: application-access       partition: 111  leader: -1      replicas:
> > 4,6,1 isr: 1
> >
> > topic: bill-generated   partition: 3    leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: bill-generated   partition: 33   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: bill-generated   partition: 63   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: bill-generated   partition: 93   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: bill-generated   partition: 123  leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: collected-event  partition: 29   leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: collected-event  partition: 59   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: collected-event  partition: 89   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: collected-event  partition: 119  leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: customer-cues    partition: 27   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: customer-cues    partition: 57   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: customer-cues    partition: 87   leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: customer-cues    partition: 117  leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: dish-promo-application-access    partition: 23   leader: -1
> > replicas: 4,6,1 isr:
> >
> > topic: dish-promo-application-access    partition: 53   leader: -1
> > replicas: 4,6,1 isr:
> >
> > topic: dish-promo-application-access    partition: 83   leader: -1
> > replicas: 4,6,1 isr:
> >
> > topic: dish-promo-application-access    partition: 113  leader: -1
> > replicas: 4,6,1 isr:
> >
> > topic: event-response   partition: 2    leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: event-response   partition: 32   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: event-response   partition: 62   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: event-response   partition: 92   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: event-response   partition: 122  leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: leads-service    partition: 24   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: leads-service    partition: 54   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: leads-service    partition: 84   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: leads-service    partition: 114  leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: logprod_v3       partition: 3    leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: logprod_v3       partition: 33   leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: logprod_v3       partition: 63   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: logprod_v3       partition: 93   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: logprod_v3       partition: 123  leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: online-account-registration-attempted    partition: 21   leader:
> > -1      replicas: 4,6,1 isr:
> >
> > topic: online-account-registration-attempted    partition: 51   leader:
> > -1      replicas: 4,6,1 isr: 1
> >
> > topic: online-account-registration-attempted    partition: 81   leader:
> > -1      replicas: 4,6,1 isr:
> >
> > topic: online-account-registration-attempted    partition: 111  leader:
> > -1      replicas: 4,6,1 isr:
> >
> > topic: order-cancelled  partition: 29   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: order-cancelled  partition: 59   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: order-cancelled  partition: 89   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: order-cancelled  partition: 119  leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: order-completed  partition: 24   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: order-completed  partition: 54   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: order-completed  partition: 84   leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: order-completed  partition: 114  leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: order-created    partition: 25   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: order-created    partition: 55   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: order-created    partition: 85   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: order-created    partition: 115  leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: order-modified   partition: 8    leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: order-modified   partition: 38   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: order-modified   partition: 68   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: order-modified   partition: 98   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: order-modified   partition: 128  leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: order-request    partition: 24   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: order-request    partition: 54   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: order-request    partition: 84   leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: order-request    partition: 114  leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: order-response   partition: 27   leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: order-response   partition: 57   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: order-response   partition: 87   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: order-response   partition: 117  leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: outbound-call-attempted  partition: 13   leader: -1      replicas:
> > 4,6,1 isr:
> >
> > topic: outbound-call-attempted  partition: 43   leader: -1      replicas:
> > 4,6,1 isr: 1
> >
> > topic: outbound-call-attempted  partition: 73   leader: -1      replicas:
> > 4,6,1 isr: 1
> >
> > topic: outbound-call-attempted  partition: 103  leader: -1      replicas:
> > 4,6,1 isr:
> >
> > topic: outbound-communications  partition: 4    leader: -1      replicas:
> > 4,6,1 isr:
> >
> > topic: outbound-communications  partition: 34   leader: -1      replicas:
> > 4,6,1 isr:
> >
> > topic: outbound-communications  partition: 64   leader: -1      replicas:
> > 4,6,1 isr:
> >
> > topic: outbound-communications  partition: 94   leader: -1      replicas:
> > 4,6,1 isr: 1
> >
> > topic: outbound-communications  partition: 124  leader: -1      replicas:
> > 4,6,1 isr: 1
> >
> > topic: postal-mail-undeliverable        partition: 15   leader: -1
> > replicas: 4,6,1 isr: 1
> >
> > topic: postal-mail-undeliverable        partition: 45   leader: -1
> > replicas: 4,6,1 isr:
> >
> > topic: postal-mail-undeliverable        partition: 75   leader: -1
> > replicas: 4,6,1 isr:
> >
> > topic: postal-mail-undeliverable        partition: 105  leader: -1
> > replicas: 4,6,1 isr:
> >
> > topic: rain-burn-in     partition: 4    leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: rain-burn-in     partition: 34   leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: rain-burn-in     partition: 64   leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: rain-burn-in     partition: 94   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: rain-burn-in     partition: 124  leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: rain-enhanced    partition: 26   leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: rain-enhanced    partition: 56   leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: rain-enhanced    partition: 86   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: rain-enhanced    partition: 116  leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: rain-listener    partition: 23   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: rain-listener    partition: 53   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: rain-listener    partition: 83   leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: rain-listener    partition: 113  leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: rain-load-test   partition: 8    leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: rain-load-test   partition: 38   leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: rain-load-test   partition: 68   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: rain-load-test   partition: 98   leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: rain-load-test   partition: 128  leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: submit-agreement partition: 2    leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: submit-agreement partition: 32   leader: -1      replicas: 4,6,1
> > isr: 1
> >
> > topic: submit-agreement partition: 62   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: submit-agreement partition: 92   leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: submit-agreement partition: 122  leader: -1      replicas: 4,6,1
> > isr:
> >
> > topic: threshold-exceeded       partition: 14   leader: -1      replicas:
> > 4,6,1 isr:
> >
> > topic: threshold-exceeded       partition: 44   leader: -1      replicas:
> > 4,6,1 isr:
> >
> > topic: threshold-exceeded       partition: 74   leader: -1      replicas:
> > 4,6,1 isr:
> >
> > topic: threshold-exceeded       partition: 104  leader: -1      replicas:
> > 4,6,1 isr: 1
> >
> >
> >
> > *Andy Nielsen*
> >
> > *Middleware Application Admin*
> >
> > *303-723-2347<tel:303-723-2347> <303-723-2347<tel:303-723-2347>>*
> >
> > *cell:720-971-2856<tel:720-971-2856> <720-971-2856<tel:720-971-2856>>*
> >
> >
> >
>
>
>

Reply via email to