I would love to help you guys to make Kafka best in Pub/Sub, will continue 
doing that whenever I can.

Do we have 0.8.1.2 release tag  or should we apply patch on top of 0.8.1.1 tag 
because we need this KAFKA-1382 JIRA ?.

Balaji

From: Neha Narkhede [mailto:neha.narkh...@gmail.com]
Sent: Monday, September 29, 2014 5:21 PM
To: Seshadri, Balaji
Cc: users@kafka.apache.org
Subject: Re: BadVersion state in Kafka Logs

It is difficult to predict an exact date. Though all the discussions of the 
progress and ETA are on the mailing list. You can follow the discussions to 
know the details and/or offer to help out on the outstanding issues.

On Mon, Sep 29, 2014 at 3:48 PM, Seshadri, Balaji 
<balaji.sesha...@dish.com<mailto:balaji.sesha...@dish.com>> wrote:
Neha,

Do you know the date in Oct when 0.8.2 is going to be out ?.

Thanks,

Balaji

From: Neha Narkhede 
[mailto:neha.narkh...@gmail.com<mailto:neha.narkh...@gmail.com>]
Sent: Thursday, September 25, 2014 1:08 PM
To: Seshadri, Balaji
Cc: users@kafka.apache.org<mailto:users@kafka.apache.org>

Subject: Re: BadVersion state in Kafka Logs

We are close to the release. I'd probably expect 0.8.2 sometime in October.

On Thu, Sep 25, 2014 at 10:37 AM, Seshadri, Balaji 
<balaji.sesha...@dish.com<mailto:balaji.sesha...@dish.com>> wrote:
Hi Neha,

Do you know when are you guys releasing 0.8.2 ?.

Thanks,

Balaji

-----Original Message-----
From: Seshadri, Balaji 
[mailto:balaji.sesha...@dish.com<mailto:balaji.sesha...@dish.com>]
Sent: Thursday, September 25, 2014 9:41 AM
To: users@kafka.apache.org<mailto:users@kafka.apache.org>
Subject: RE: BadVersion state in Kafka Logs

Thanks for the replay.

Please let me know if we can use trunk as 0.8.2 is not yet released.

Balaji
________________________________________
From: Neha Narkhede [neha.narkh...@gmail.com<mailto:neha.narkh...@gmail.com>]
Sent: Wednesday, September 24, 2014 6:32 PM
To: users@kafka.apache.org<mailto:users@kafka.apache.org>
Subject: Re: BadVersion state in Kafka Logs

From the logs you've attached, my guess is it's most likely due to KAFKA-1382.

Thanks,
Neha

On Wed, Sep 24, 2014 at 10:48 AM, Seshadri, Balaji 
<balaji.sesha...@dish.com<mailto:balaji.sesha...@dish.com>
> wrote:

> Hi,
>
>
>
> We got the below error in our logs and our consumers stopped consuming
> any data ?.It worked only after restart.
>
>
>
> We would like to confirm that it's because we are running with
> 0.8-beta version and not 0.8 release version to convince "THE MGMT" guys.
>
>
>
> Please let me know if it's this KAFKA-1382 causing the issue.
>
>
>
> Thanks,
>
>
>
> Balaji
>
>
>
> *From:* Gulia, Vikram
> *Sent:* Wednesday, September 24, 2014 8:43 AM
> *To:* Sharma, Navdeep; #IT-MAD DES; #IT-MAA
> *Cc:* Alam, Mohammad Shah
> *Subject:* RE: 9/23 prod issue - offline kafka partitions.
>
>
>
> Adding full MAA distro.
>
>
>
> DES Offshore looked in to the logs on kafka servers and seems like the
> issue we encountered yesterday may be described in these threads,
> please have a look -
>
>
>
> http://permalink.gmane.org/gmane.comp.apache.kafka.user/1904
>
>
>
> https://issues.apache.org/jira/browse/KAFKA-1382 (it describes the
> fix/patch which is available in 0.8.1.2/0.8.2<http://0.8.1.2/0.8.2>)
>
>
>
> Thank You,
>
> Vikram Gulia
>
>
>
> *From:* Sharma, Navdeep
> *Sent:* Wednesday, September 24, 2014 6:53 AM
> *To:* Gulia, Vikram; #IT-MAD DES
> *Cc:* #IT-MAA Offshore; Alam, Mohammad Shah
> *Subject:* RE: 9/23 prod issue - offline kafka partitions.
>
>
>
> Hi Vikram,
>
>
>
> We analyzed  below mentioned issue with MAA-Offshore (Abhishek) and
> found that the error occurred only on 23 Sept. This is  not historical
> as we checked last 4 days logs.
>
>
>
> It looks like that consumer got stopped on September 22 2014 for Linux
> patching activty.MAA started consumer September  23 2014 at 1:00 AM.
>
>
>
> *Issue *in  server log   *"* *BadVersion for
> /brokers/topics/rain-burn-in/partitions/121/state"*  but it is not
> present in previous 4 days logs.
>
> More detail of this error can be found at-
>
> http://permalink.gmane.org/gmane.comp.apache.kafka.user/1904
>
>
>
> We are not sure about data loss in this scenario and working on this.
>
>
>
>
>
>
>
> Let us know if any concerns.
>
>
>
> [image: cid:image001.gif@01CF7B0A.03F21580]
>
> Navdeep Sharma
> Developer - offshore,  Middleware Applications & Development o India:
> 0120-4532000 - 2234
> c: +91-9911698102<tel:%2B91-9911698102>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
> *From:* Gulia, Vikram
> *Sent:* Tuesday, September 23, 2014 6:17 PM
> *To:* #IT-MAD DES
> *Subject:* FW: 9/23 prod issue - offline kafka partitions.
>
>
>
> DES Offshore dev,
>
>
>
> Please work with MAA offshore to monitor the kafka broker as we had
> this incident where lot of partitions went offline around 1.45 PM MST
> and MAA has to restart the kafka servers. We may have lost messages
> and we need to see if there is a way to figure out what was the impact.
>
>
>
> Also, check the logs for kafka servers and see if we can figure out
> why did partitions go offline or are un-available? Let us know if you
> find anything relevant.
>
>
>
> Thank You,
>
> Vikram Gulia
>
>
>
> *From:* Nielsen, Andy
> *Sent:* Tuesday, September 23, 2014 5:04 PM
> *To:* #IT-MAD DES; Gulia, Vikram
> *Cc:* #IT-MAA
> *Subject:* 9/23 prod issue - offline kafka partitions.
>
>
>
> desadmin@pc1mwdpl01 ~/bin $ ./kafka.sh topic --unavailable-partitions
>
> topic: account-access   partition: 21   leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: account-access   partition: 51   leader: -1      replicas: 4,6,1
> isr:
>
> topic: account-access   partition: 81   leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: account-access   partition: 111  leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: account-activated        partition: 13   leader: -1      replicas:
> 4,6,1 isr:
>
> topic: account-activated        partition: 43   leader: -1      replicas:
> 4,6,1 isr:
>
> topic: account-activated        partition: 73   leader: -1      replicas:
> 4,6,1 isr:
>
> topic: account-activated        partition: 103  leader: -1      replicas:
> 4,6,1 isr: 1
>
> topic: account-adjustment-issued        partition: 27   leader: -1
> replicas: 4,6,1 isr:
>
> topic: account-adjustment-issued        partition: 57   leader: -1
> replicas: 4,6,1 isr:
>
> topic: account-adjustment-issued        partition: 87   leader: -1
> replicas: 4,6,1 isr: 1
>
> topic: account-adjustment-issued        partition: 117  leader: -1
> replicas: 4,6,1 isr:
>
> topic: account-created  partition: 11   leader: -1      replicas: 4,6,1
> isr:
>
> topic: account-created  partition: 41   leader: -1      replicas: 4,6,1
> isr:
>
> topic: account-created  partition: 71   leader: -1      replicas: 4,6,1
> isr:
>
> topic: account-created  partition: 101  leader: -1      replicas: 4,6,1
> isr:
>
> topic: account-info-updated     partition: 7    leader: -1      replicas:
> 4,6,1 isr: 1
>
> topic: account-info-updated     partition: 37   leader: -1      replicas:
> 4,6,1 isr: 1
>
> topic: account-info-updated     partition: 67   leader: -1      replicas:
> 4,6,1 isr:
>
> topic: account-info-updated     partition: 97   leader: -1      replicas:
> 4,6,1 isr: 1
>
> topic: account-info-updated     partition: 127  leader: -1      replicas:
> 4,6,1 isr: 1
>
> topic: application-access       partition: 21   leader: -1      replicas:
> 4,6,1 isr: 1
>
> topic: application-access       partition: 51   leader: -1      replicas:
> 4,6,1 isr: 1
>
> topic: application-access       partition: 81   leader: -1      replicas:
> 4,6,1 isr: 1
>
> topic: application-access       partition: 111  leader: -1      replicas:
> 4,6,1 isr: 1
>
> topic: bill-generated   partition: 3    leader: -1      replicas: 4,6,1
> isr:
>
> topic: bill-generated   partition: 33   leader: -1      replicas: 4,6,1
> isr:
>
> topic: bill-generated   partition: 63   leader: -1      replicas: 4,6,1
> isr:
>
> topic: bill-generated   partition: 93   leader: -1      replicas: 4,6,1
> isr:
>
> topic: bill-generated   partition: 123  leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: collected-event  partition: 29   leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: collected-event  partition: 59   leader: -1      replicas: 4,6,1
> isr:
>
> topic: collected-event  partition: 89   leader: -1      replicas: 4,6,1
> isr:
>
> topic: collected-event  partition: 119  leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: customer-cues    partition: 27   leader: -1      replicas: 4,6,1
> isr:
>
> topic: customer-cues    partition: 57   leader: -1      replicas: 4,6,1
> isr:
>
> topic: customer-cues    partition: 87   leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: customer-cues    partition: 117  leader: -1      replicas: 4,6,1
> isr:
>
> topic: dish-promo-application-access    partition: 23   leader: -1
> replicas: 4,6,1 isr:
>
> topic: dish-promo-application-access    partition: 53   leader: -1
> replicas: 4,6,1 isr:
>
> topic: dish-promo-application-access    partition: 83   leader: -1
> replicas: 4,6,1 isr:
>
> topic: dish-promo-application-access    partition: 113  leader: -1
> replicas: 4,6,1 isr:
>
> topic: event-response   partition: 2    leader: -1      replicas: 4,6,1
> isr:
>
> topic: event-response   partition: 32   leader: -1      replicas: 4,6,1
> isr:
>
> topic: event-response   partition: 62   leader: -1      replicas: 4,6,1
> isr:
>
> topic: event-response   partition: 92   leader: -1      replicas: 4,6,1
> isr:
>
> topic: event-response   partition: 122  leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: leads-service    partition: 24   leader: -1      replicas: 4,6,1
> isr:
>
> topic: leads-service    partition: 54   leader: -1      replicas: 4,6,1
> isr:
>
> topic: leads-service    partition: 84   leader: -1      replicas: 4,6,1
> isr:
>
> topic: leads-service    partition: 114  leader: -1      replicas: 4,6,1
> isr:
>
> topic: logprod_v3       partition: 3    leader: -1      replicas: 4,6,1
> isr:
>
> topic: logprod_v3       partition: 33   leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: logprod_v3       partition: 63   leader: -1      replicas: 4,6,1
> isr:
>
> topic: logprod_v3       partition: 93   leader: -1      replicas: 4,6,1
> isr:
>
> topic: logprod_v3       partition: 123  leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: online-account-registration-attempted    partition: 21   leader:
> -1      replicas: 4,6,1 isr:
>
> topic: online-account-registration-attempted    partition: 51   leader:
> -1      replicas: 4,6,1 isr: 1
>
> topic: online-account-registration-attempted    partition: 81   leader:
> -1      replicas: 4,6,1 isr:
>
> topic: online-account-registration-attempted    partition: 111  leader:
> -1      replicas: 4,6,1 isr:
>
> topic: order-cancelled  partition: 29   leader: -1      replicas: 4,6,1
> isr:
>
> topic: order-cancelled  partition: 59   leader: -1      replicas: 4,6,1
> isr:
>
> topic: order-cancelled  partition: 89   leader: -1      replicas: 4,6,1
> isr:
>
> topic: order-cancelled  partition: 119  leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: order-completed  partition: 24   leader: -1      replicas: 4,6,1
> isr:
>
> topic: order-completed  partition: 54   leader: -1      replicas: 4,6,1
> isr:
>
> topic: order-completed  partition: 84   leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: order-completed  partition: 114  leader: -1      replicas: 4,6,1
> isr:
>
> topic: order-created    partition: 25   leader: -1      replicas: 4,6,1
> isr:
>
> topic: order-created    partition: 55   leader: -1      replicas: 4,6,1
> isr:
>
> topic: order-created    partition: 85   leader: -1      replicas: 4,6,1
> isr:
>
> topic: order-created    partition: 115  leader: -1      replicas: 4,6,1
> isr:
>
> topic: order-modified   partition: 8    leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: order-modified   partition: 38   leader: -1      replicas: 4,6,1
> isr:
>
> topic: order-modified   partition: 68   leader: -1      replicas: 4,6,1
> isr:
>
> topic: order-modified   partition: 98   leader: -1      replicas: 4,6,1
> isr:
>
> topic: order-modified   partition: 128  leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: order-request    partition: 24   leader: -1      replicas: 4,6,1
> isr:
>
> topic: order-request    partition: 54   leader: -1      replicas: 4,6,1
> isr:
>
> topic: order-request    partition: 84   leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: order-request    partition: 114  leader: -1      replicas: 4,6,1
> isr:
>
> topic: order-response   partition: 27   leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: order-response   partition: 57   leader: -1      replicas: 4,6,1
> isr:
>
> topic: order-response   partition: 87   leader: -1      replicas: 4,6,1
> isr:
>
> topic: order-response   partition: 117  leader: -1      replicas: 4,6,1
> isr:
>
> topic: outbound-call-attempted  partition: 13   leader: -1      replicas:
> 4,6,1 isr:
>
> topic: outbound-call-attempted  partition: 43   leader: -1      replicas:
> 4,6,1 isr: 1
>
> topic: outbound-call-attempted  partition: 73   leader: -1      replicas:
> 4,6,1 isr: 1
>
> topic: outbound-call-attempted  partition: 103  leader: -1      replicas:
> 4,6,1 isr:
>
> topic: outbound-communications  partition: 4    leader: -1      replicas:
> 4,6,1 isr:
>
> topic: outbound-communications  partition: 34   leader: -1      replicas:
> 4,6,1 isr:
>
> topic: outbound-communications  partition: 64   leader: -1      replicas:
> 4,6,1 isr:
>
> topic: outbound-communications  partition: 94   leader: -1      replicas:
> 4,6,1 isr: 1
>
> topic: outbound-communications  partition: 124  leader: -1      replicas:
> 4,6,1 isr: 1
>
> topic: postal-mail-undeliverable        partition: 15   leader: -1
> replicas: 4,6,1 isr: 1
>
> topic: postal-mail-undeliverable        partition: 45   leader: -1
> replicas: 4,6,1 isr:
>
> topic: postal-mail-undeliverable        partition: 75   leader: -1
> replicas: 4,6,1 isr:
>
> topic: postal-mail-undeliverable        partition: 105  leader: -1
> replicas: 4,6,1 isr:
>
> topic: rain-burn-in     partition: 4    leader: -1      replicas: 4,6,1
> isr:
>
> topic: rain-burn-in     partition: 34   leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: rain-burn-in     partition: 64   leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: rain-burn-in     partition: 94   leader: -1      replicas: 4,6,1
> isr:
>
> topic: rain-burn-in     partition: 124  leader: -1      replicas: 4,6,1
> isr:
>
> topic: rain-enhanced    partition: 26   leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: rain-enhanced    partition: 56   leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: rain-enhanced    partition: 86   leader: -1      replicas: 4,6,1
> isr:
>
> topic: rain-enhanced    partition: 116  leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: rain-listener    partition: 23   leader: -1      replicas: 4,6,1
> isr:
>
> topic: rain-listener    partition: 53   leader: -1      replicas: 4,6,1
> isr:
>
> topic: rain-listener    partition: 83   leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: rain-listener    partition: 113  leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: rain-load-test   partition: 8    leader: -1      replicas: 4,6,1
> isr:
>
> topic: rain-load-test   partition: 38   leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: rain-load-test   partition: 68   leader: -1      replicas: 4,6,1
> isr:
>
> topic: rain-load-test   partition: 98   leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: rain-load-test   partition: 128  leader: -1      replicas: 4,6,1
> isr:
>
> topic: submit-agreement partition: 2    leader: -1      replicas: 4,6,1
> isr:
>
> topic: submit-agreement partition: 32   leader: -1      replicas: 4,6,1
> isr: 1
>
> topic: submit-agreement partition: 62   leader: -1      replicas: 4,6,1
> isr:
>
> topic: submit-agreement partition: 92   leader: -1      replicas: 4,6,1
> isr:
>
> topic: submit-agreement partition: 122  leader: -1      replicas: 4,6,1
> isr:
>
> topic: threshold-exceeded       partition: 14   leader: -1      replicas:
> 4,6,1 isr:
>
> topic: threshold-exceeded       partition: 44   leader: -1      replicas:
> 4,6,1 isr:
>
> topic: threshold-exceeded       partition: 74   leader: -1      replicas:
> 4,6,1 isr:
>
> topic: threshold-exceeded       partition: 104  leader: -1      replicas:
> 4,6,1 isr: 1
>
>
>
> *Andy Nielsen*
>
> *Middleware Application Admin*
>
> *303-723-2347<tel:303-723-2347> <303-723-2347<tel:303-723-2347>>*
>
> *cell:720-971-2856<tel:720-971-2856> <720-971-2856<tel:720-971-2856>>*
>
>
>


Reply via email to