Hi Harsh/Ismael,

Any suggestions or inputs for the above issue?

When i run the producer client, I still get this error

./kafka-console-producer.sh --broker-list hostname:9094 --topic topic3


*[2016-01-05 10:16:20,272] ERROR Error when sending message to topic test
with key: null, value: 5 bytes with error: Failed to update metadata after
60000 ms.
(org.apache.kafka.clients.producer.internals.ErrorLoggingCallback)*

Also, i can see the below error in controller.log

*[2016-01-20 09:39:33,408] DEBUG [Controller 0]: preferred replicas by
broker Map(1 -> Map([topic3,0] -> List(1, 0)), 0 -> Map([topic3,1] ->
List(0, 1), [topic2,0] -> List(0), [topic1,0] -> List(0)))
(kafka.controller.KafkaController)*
*[2016-01-20 09:39:33,408] DEBUG [Controller 0]: topics not in preferred
replica Map() (kafka.controller.KafkaController)*
*[2016-01-20 09:39:33,408] TRACE [Controller 0]: leader imbalance ratio for
broker 1 is 0.000000 (kafka.controller.KafkaController)*
*[2016-01-20 09:39:33,408] DEBUG [Controller 0]: topics not in preferred
replica Map() (kafka.controller.KafkaController)*
*[2016-01-20 09:39:33,409] TRACE [Controller 0]: leader imbalance ratio for
broker 0 is 0.000000 (kafka.controller.KafkaController)*


Tried reinstalling kafka, but no luck:(


Checked telnet also, I am able to connect to that port.
[root@blrd-cmgvapp46 logs]# telnet hostname 9094
Trying 172.31.31.186...
Connected to hostname (172.31.31.186).
Escape character is '^]'.

I can see the topic is created properly.

[root@hostname bin]# ./kafka-topics.sh --describe --zookeeper hostname:2181
--topic topic3
Topic:topic3    PartitionCount:2        ReplicationFactor:2     Configs:
        Topic: topic3   Partition: 0    Leader: 1       Replicas: 1,0
Isr: 1,0
        Topic: topic3   Partition: 1    Leader: 0       Replicas: 0,1
Isr: 0,1


Thanks in advance,


On Tue, Jan 5, 2016 at 3:17 PM, prabhu v <prabhuvrajp...@gmail.com> wrote:

> Hi Harsha,
>
> This is my Kafka_server_jaas.config file. This is passed as JVM param to
> the Kafka broker while start up.
>
> =============
> KafkaServer {
>     com.sun.security.auth.module.Krb5LoginModule required
>       useKeyTab=true
>        storeKey=true
>       serviceName="kafka"
>        keyTab="/etc/security/keytabs/kafka1.keytab"
>         useTicketCache=true
>         principal="kafka/hostname@realmname";
> };
>
> zkclient{
>
> com.sun.security.auth.module.Krb5LoginModule required
>       useKeyTab=true
>        storeKey=true
>       serviceName="zookeeper"
>        keyTab="/etc/security/keytabs/kafka1.keytab"
>         useTicketCache=true
>         principal="kafka@realmname";
>
> };
> =============
>
> Note: For security reasons, changed my original FQDN to hostname and
> original realm name to realm name in the below output.
>
> I am able to view the ticket using klist command as well. Please find
> below output.
>
> [root@localhost config]# kinit -k -t /etc/security/keytabs/kafka1.keytab
> kafka/hostname@realmname
> [root@localhost config]# klist
> Ticket cache: FILE:/tmp/krb5cc_0
> Default principal: kafka/hostname@realmname
>
> Valid starting     Expires            Service principal
> 01/05/16 08:14:28  01/06/16 08:14:28  krbtgt/realm@realm
>         renew until 01/05/16 08:14:28
>
>
>
>
>
>
> For(topics,producer and consumer) clients, I am using the below JAAS
> Config:
>
> =============
>
> Client {
> com.sun.security.auth.module.Krb5LoginModule required
> useKeyTab=true
> keyTab="/etc/security/keytabs/kafka_client.keytab"
> storeKey=true
> useTicketCache=true
> serviceName="kafka"
> principal="kafkaclient/hostname@realmname";
> };
>
> =============
>
> I am able to view the ticket using klist command as well. Please find
> below output.
>
> [root@localhost config]# kinit -k -t
> /etc/security/keytabs/kafka_client.keytab kafkaclient/hostname@realmname
> [root@localhost config]# klist
> Ticket cache: FILE:/tmp/krb5cc_0
> Default principal: kafkaclient/hostname@realmname
>
> Valid starting     Expires            Service principal
> 01/05/16 08:14:28  01/06/16 08:14:28  krbtgt/realm@realm
>         renew until 01/05/16 08:14:28
>
> Error when running producer client:
>
> ./kafka-console-producer.sh --broker-list hostname:9095 --topic test
>
>
> [2016-01-05 10:16:20,272] ERROR Error when sending message to topic test
> with key: null, value: 5 bytes with error: Failed to update metadata after
> 60000 ms. (org.apache.kafka.clients.producer.internals.ErrorLoggingCallback)
>
> Error when running topics.sh:
>
> [root@localhost bin]# ./kafka-topics.sh --list --zookeeper hostname:2181
> [2015-12-28 12:41:32,589] WARN SASL configuration failed:
> javax.security.auth.login.LoginException: No key to store Will continue
> connection to Zookeeper server without SASL authentication, if Zookeeper
> server allows it. (org.apache.zookeeper.ClientCnxn)
> ^Z
>
> Please let me know if i am missing anything.
>
>
>
>
> Thanks,
> Prabhu
>
>
>
>
> On Wed, Dec 30, 2015 at 9:28 PM, Harsha <ka...@harsha.io> wrote:
>
>> can you add your jass file details. Your jaas file might have
>> useTicketCache=true and storeKey=true as well
>> example of KafkaServer jass file
>>
>> KafkaServer {
>>
>> com.sun.security.auth.module.Krb5LoginModule required
>>
>> useKeyTab=true
>>
>> storeKey=true
>>
>> serviceName="kafka"
>>
>> keyTab="/vagrant/keytabs/kafka1.keytab"
>>
>> principal="kafka/kafka1.witzend....@witzend.com";
>> };
>>
>> and KafkaClient
>> KafkaClient {
>>
>> com.sun.security.auth.module.Krb5LoginModule required
>>
>> useTicketCache=true
>>
>> serviceName="kafka";
>>
>> };
>>
>> On Wed, Dec 30, 2015, at 03:10 AM, prabhu v wrote:
>>
>> Hi Harsha,
>>
>> I have used the Fully qualified domain name. Just for security concerns,
>> Before sending this mail,i have replaced our FQDN hostname to localhost.
>>
>> yes, i have tried KINIT and I am able to view the tickets using klist
>> command as well.
>>
>> Thanks,
>> Prabhu
>>
>> On Wed, Dec 30, 2015 at 11:27 AM, Harsha <ka...@harsha.io> wrote:
>>
>> Prabhu,
>>            When using SASL/kerberos always make sure you give FQDN of
>>            the hostname . In your command you are using --zookeeper
>>            localhost:2181 and make sure you change that hostname.
>>
>> "avax.security.auth.login.LoginException: No key to store Will continue
>> > connection to Zookeeper server without SASL authentication, if
>> Zookeeper"
>>
>> did you try  kinit with that keytab at the command line.
>>
>> -Harsha
>> On Mon, Dec 28, 2015, at 04:07 AM, prabhu v wrote:
>> > Thanks for the input Ismael.
>> >
>> > I will try and let you know.
>> >
>> > Also need your valuable inputs for the below issue:)
>> >
>> > i am not able to run kafka-topics.sh(0.9.0.0 version)
>> >
>> > [root@localhost bin]# ./kafka-topics.sh --list --zookeeper
>> localhost:2181
>> > [2015-12-28 12:41:32,589] WARN SASL configuration failed:
>> > javax.security.auth.login.LoginException: No key to store Will continue
>> > connection to Zookeeper server without SASL authentication, if Zookeeper
>> > server allows it. (org.apache.zookeeper.ClientCnxn)
>> > ^Z
>> >
>> > I am sure the key is present in its keytab file ( I have cross verified
>> > using kinit command as well).
>> >
>> > Am i missing anything while calling the kafka-topics.sh??
>> >
>> >
>> >
>> > On Mon, Dec 28, 2015 at 3:53 PM, Ismael Juma <isma...@gmail.com> wrote:
>> >
>> > > Hi Prabhu,
>> > >
>> > > kafka-console-consumer.sh uses the old consumer by default, but only
>> the
>> > > new consumer supports security. Use --new-consumer to change this.
>> > >
>> > > Hope this helps.
>> > >
>> > > Ismael
>> > > On 28 Dec 2015 05:48, "prabhu v" <prabhuvrajp...@gmail.com> wrote:
>> > >
>> > > > Hi Experts,
>> > > >
>> > > > I am getting the below error when running the consumer
>> > > > "kafka-console-consumer.sh" .
>> > > >
>> > > > I am using the new version 0.9.0.1.
>> > > > Topic name: test
>> > > >
>> > > >
>> > > > [2015-12-28 06:13:34,409] WARN
>> > > >
>> > > >
>> > >
>> [console-consumer-61657_localhost-1451283204993-5512891d-leader-finder-thread],
>> > > > Failed to find leader for Set([test,0])
>> > > > (kafka.consumer.ConsumerFetcherManager$LeaderFinderThread)
>> > > > kafka.common.BrokerEndPointNotAvailableException: End point
>> PLAINTEXT not
>> > > > found for broker 0
>> > > >         at kafka.cluster.Broker.getBrokerEndPoint(Broker.scala:136)
>> > > >
>> > > >
>> > > > Please find the current configuration below.
>> > > >
>> > > > Configuration:
>> > > >
>> > > >
>> > > > [root@localhost config]# grep -v "^#" consumer.properties
>> > > > zookeeper.connect=localhost:2181
>> > > > zookeeper.connection.timeout.ms=60000
>> > > > group.id=test-consumer-group
>> > > > security.protocol=SASL_PLAINTEXT
>> > > > sasl.kerberos.service.name="kafka"
>> > > >
>> > > >
>> > > > [root@localhost config]# grep -v "^#" producer.properties
>> > > > metadata.broker.list=localhost:9094,localhost:9095
>> > > > producer.type=sync
>> > > > compression.codec=none
>> > > > serializer.class=kafka.serializer.DefaultEncoder
>> > > > security.protocol=SASL_PLAINTEXT
>> > > > sasl.kerberos.service.name="kafka"
>> > > >
>> > > > [root@localhost config]# grep -v "^#" server1.properties
>> > > >
>> > > > broker.id=0
>> > > > listeners=SASL_PLAINTEXT://localhost:9094
>> > > > delete.topic.enable=true
>> > > > num.network.threads=3
>> > > > num.io.threads=8
>> > > > socket.send.buffer.bytes=102400
>> > > > socket.receive.buffer.bytes=102400
>> > > > socket.request.max.bytes=104857600
>> > > > log.dirs=/data/kafka_2.11-0.9.0.0/kafka-logs
>> > > > num.partitions=1
>> > > > num.recovery.threads.per.data.dir=1
>> > > > log.retention.hours=168
>> > > > log.segment.bytes=1073741824
>> > > > log.retention.check.interval.ms=300000
>> > > > log.cleaner.enable=false
>> > > > zookeeper.connect=localhost:2181
>> > > > zookeeper.connection.timeout.ms=60000
>> > > > inter.broker.protocol.version=0.9.0.0
>> > > > security.inter.broker.protocol=SASL_PLAINTEXT
>> > > > allow.everyone.if.no.acl.found=true
>> > > >
>> > > >
>> > > > [root@localhost config]# grep -v "^#" server4.properties
>> > > > broker.id=1
>> > > > listeners=SASL_PLAINTEXT://localhost:9095
>> > > > delete.topic.enable=true
>> > > > num.network.threads=3
>> > > > num.io.threads=8
>> > > > socket.send.buffer.bytes=102400
>> > > > socket.receive.buffer.bytes=102400
>> > > > socket.request.max.bytes=104857600
>> > > > log.dirs=/data/kafka_2.11-0.9.0.0/kafka-logs-1
>> > > > num.partitions=1
>> > > > num.recovery.threads.per.data.dir=1
>> > > > log.retention.hours=168
>> > > > log.segment.bytes=1073741824
>> > > > log.retention.check.interval.ms=300000
>> > > > log.cleaner.enable=false
>> > > > zookeeper.connect=localhost:2181
>> > > > zookeeper.connection.timeout.ms=60000
>> > > > inter.broker.protocol.version=0.9.0.0
>> > > > security.inter.broker.protocol=SASL_PLAINTEXT
>> > > > zookeeper.sasl.client=zkclient
>> > > >
>> > > > [root@localhost config]# grep -v "^#" zookeeper.properties
>> > > > dataDir=/data/kafka_2.11-0.9.0.0/zookeeper
>> > > > clientPort=2181
>> > > > maxClientCnxns=0
>> > > > requireClientAuthScheme=sasl
>> > > >
>> > >
>> authProvider.1=org.apache.zookeeper.server.auth.SASLAuthenticationProvider
>> > > > jaasLoginRenew=3600000
>> > > >
>> > > >
>> > > > Need your valuable inputs on this issue.
>> > > > --
>> > > > Regards,
>> > > >
>> > > > Prabhu.V
>> > > >
>> > >
>> >
>> >
>> >
>> > --
>> > Regards,
>> >
>> > Prabhu.V
>>
>>
>>
>>
>> --
>> Regards,
>>
>> Prabhu.V
>>
>>
>>
>>
>
>
>
> --
> Regards,
>
> Prabhu.V
>
>



-- 
Regards,

Prabhu.V

Reply via email to