Re: Can I make Nimbus failsafe

2014-09-16 Thread Parth Brahmbhatt
I am working on this right now , can’t promise the dates because I have some 
other tasks as well. The general approach is already described in the JIRA. You 
basically need 3 things:

Leader election between all nimbus hosts. (zookeeper is what I am going with), 
checks on nimbus background threads and state changing APIS to ensure only 
leader is changing states.
Some way to distribute the local nimbus data directory to all the other nimbus 
and keep it consistent. Simplest way is to not have a local directory but 
instead use something like HDFS, if you can take that kind of dependency.
Nimbus discovery, could be done either by adding/modifying current rest/thrift 
APIS or by allowing supervisors or clients to read from zookeeper. 

I have some code on leader election part with preliminary testing and unit-test 
here. This is by no means ready for anyone’s use but if you want to take a look 
to get an idea you can take a look.

Thanks
Parth

On Sep 16, 2014, at 6:04 AM, M.Tarkeshwar Rao tarkeshwa...@gmail.com wrote:

 something


-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.


Re: How Do Workers Connect To Nimbus

2014-09-08 Thread Parth Brahmbhatt
One thing that is generally needed on EC2 hosts is security group setting 

http://docs.aws.amazon.com/gettingstarted/latest/computebasics/getting-started-security-group.html

If you can not ping the host I am guessing this might be one of the things that 
you need to setup correctly.

On Sep 8, 2014, at 3:34 PM, Stephen Hartzell hartzell.step...@gmail.com wrote:

 cannot


-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.


Re: How Do Workers Connect To Nimbus

2014-09-08 Thread Parth Brahmbhatt
The worker connects to the thrift port and not the ui port. You need to open 
port 6627 or whatever is the value being set in storm.yaml using  property 
“nimbus.thrift.port”.

Based on the configuration that you have pointed so far it seems your nimbus 
host has nimbus,ui,supervisor working because you actually have zookeeper 
running locally on that host. As Harsha pointed out you need to change it to a 
value that is the public ip instead of loopback interface.

Thanks
Parth

On Sep 8, 2014, at 3:42 PM, Harsha st...@harsha.io wrote:

 storm.zookeeper.servers:
  - 127.0.0.1
 nimbus.host: 127.0.0.1 ( 127.0.0.1 causes to bind a loopback interface , 
 instead either use your public ip or 0.0.0.0)
 storm.local.dir: /tmp/storm ( I recommend this to move to a different folder 
 probably /home/storm, /tmp/storm will get deleted if your machine is 
 restarted)
  
 make sure you zookeeper is also listening in 0.0.0.0 or public ip not 
 127.0.0.1.
  
 No, I cannot ping my host which has a public ip address of 54.68.149.181
 you are not able to reach this ip form worker node but able to access the UI 
 using it?
 -Harsha
  
 On Mon, Sep 8, 2014, at 03:34 PM, Stephen Hartzell wrote:
 Harsha, 
  
   The storm.yaml on the host machine looks like this:
  
 storm.zookeeper.servers:
  - 127.0.0.1
  
  
 nimbus.host: 127.0.0.1
  
 storm.local.dir: /tmp/storm
  
  
   The storm.yaml on the worker machine looks like this: 
  
 storm.zookeeper.servers:
  - 54.68.149.181
  
  
 nimbus.host: 54.68.149.181
  
 storm.local.dir: /tmp/storm
  
 No, I cannot ping my host which has a public ip address of 54.68.149.181 
 although I can connect to the UI web page when it is hosted. I don't know 
 how I would go about connecting to zookeeper on the nimbus host.
 -Thanks, Stephen
  
  
 On Mon, Sep 8, 2014 at 6:28 PM, Harsha st...@harsha.io wrote:
 
 There aren't any errors in worker machine supervisor logs. Are you using the 
 same storm.yaml for both the machines and also are you able to ping your 
 nimbus host or connect to zookeeper on nimbus host.
 -Harsha
 
  
  
 On Mon, Sep 8, 2014, at 03:24 PM, Stephen Hartzell wrote:
 Harsha,
  
   Thanks so much for getting back with me. I will check the logs, but I 
 don't seem to get any error messages. I have a nimbus AWS machine with 
 zookeeper on it and a worker AWS machine. 
  
 On the nimbus machine I start the zookeeper and then I run: 
  
 bin/storm nimbus 
 bin/storm supervisor 
 bin/storm ui
  
 On the worker machine I run:
 bin/storm supervisor
  
 When I go to the UI page, I only see 1 supervisor (the one on the nimbus 
 machine). So apparently, the worker machine isn't registering with the 
 nimbus machine.  
  
 On Mon, Sep 8, 2014 at 6:16 PM, Harsha st...@harsha.io wrote:
 
 Hi Stephen,
 What are the issues you are seeing.
 How do worker machines know how to connect to nimbus? Is it in the storm 
 configuration file
 
 Yes. make sure you the supervisor(worker) , nimbus nodes  are able to 
 connect to your zookeeper cluster.
 Check your logs under storm_inst/logs/ for any errors when you try to start 
 nimbus or supervisors.
 If you are installing it manually try following these steps if you are not 
 already done.
 http://www.michael-noll.com/tutorials/running-multi-node-storm-cluster/
 -Harsha
 
  
  
  
 On Mon, Sep 8, 2014, at 03:01 PM, Stephen Hartzell wrote:
 All,
  
 I would greatly appreciate any help that anyone would afford. I've been 
 trying to setup a storm cluster on AWS for a few weeks now on centOS EC2 
 machines. So far, I haven't been able to get a cluster built. I can get a 
 supervisor and nimbus to run on a single machine, but I can't figure out 
 how to get another worker to connect to nimbus. How do worker machines 
 know how to connect to nimbus? Is it in the storm configuration file? 
 I've gone through many tutorials and the official documentation, but this 
 point doesn't seem to be covered anywhere in sufficient detail for a new 
 guy like me. 
  
   Some of you may be tempted to point me toward storm-deploy, but I spent 
 four days trying to get that to work until I gave up. I'm having Issue #58 
 on github. Following the instructions exactly and other tutorials on a 
 bran new AWS machine fails. So I gave up on storm-deploy and decided to 
 try and setup a cluster manually. Thanks in advance to anyone willing to 
 offer me any inputs you can! 
  
  
  
  
  


-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. 

Re: URGENT!! Resetting KafkaSpout to fetch message from the beginning

2014-09-04 Thread Parth Brahmbhatt
Under zookeeper you should be able to find a /consumers path. I believe this is 
where the kafka consumers write their offset but I am not 100% sure.
This might be the place where all consumers storm/non-storm will be writing 
their offset so if you have non storm consumers , I would be super cautious 
before deleting the entire folder.

Thanks
Parth


On Sep 4, 2014, at 9:14 AM, Kushan Maskey kushan.mas...@mmillerassociates.com 
wrote:

 Which zookeeper data do I need to delete? Storm or Kafka?
 
 --
 Kushan Maskey
 817.403.7500
 
 
 On Thu, Sep 4, 2014 at 11:01 AM, Vikas Agarwal vi...@infoobjects.com wrote:
 I guess, you need to empty the zookeeper data directory too.
 
 
 On Thu, Sep 4, 2014 at 9:22 PM, Kushan Maskey 
 kushan.mas...@mmillerassociates.com wrote:
 I had to manually flush all the data from kafka. I did that by stopping the 
 server and deleting the entire content of kafka-logs directory. Then I 
 restarted kafka again. Then I started my storm. I get following error 
 message. Because the offset KafkaSpout is looking for is at 81573 and kafka 
 offset is at 0 right now. How do i resolve this problem?
 
 kafka.common.OffsetOutOfRangeException: Request for offset 81573 but we only 
 have log segments in the range 0 to 0.
 
 at kafka.log.Log.read(Log.scala:380)
 
 at 
 kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:530)
 
 at 
 kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:476)
 
 at 
 kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:471)
 
 at 
 scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
 
 at 
 scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
 
 at scala.collection.immutable.Map$Map1.foreach(Map.scala:105)
 
 at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
 
 at scala.collection.immutable.Map$Map1.map(Map.scala:93)
 
 at 
 kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:471)
 
 at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:437)
 
 at kafka.server.KafkaApis.handle(KafkaApis.scala:186)
 
 at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:42)
 
 at java.lang.Thread.run(Thread.java:744)
 
 [2014-09-04 15:35:25,904] ERROR [KafkaApi-0] Error when processing fetch 
 request for partition [1002,0] offset 8430 from consumer with correlation id 
 0 (kafka.server.KafkaApis)
 
 kafka.common.OffsetOutOfRangeException: Request for offset 8430 but we only 
 have log segments in the range 0 to 0.
 
 at kafka.log.Log.read(Log.scala:380)
 
 at 
 kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSet(KafkaApis.scala:530)
 
 at 
 kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:476)
 
 at 
 kafka.server.KafkaApis$$anonfun$kafka$server$KafkaApis$$readMessageSets$1.apply(KafkaApis.scala:471)
 
 at 
 scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
 
 at 
 scala.collection.TraversableLike$$anonfun$map$1.apply(TraversableLike.scala:206)
 
 at scala.collection.immutable.Map$Map1.foreach(Map.scala:105)
 
 at scala.collection.TraversableLike$class.map(TraversableLike.scala:206)
 
 at scala.collection.immutable.Map$Map1.map(Map.scala:93)
 
 at 
 kafka.server.KafkaApis.kafka$server$KafkaApis$$readMessageSets(KafkaApis.scala:471)
 
 at kafka.server.KafkaApis.handleFetchRequest(KafkaApis.scala:437)
 
 at kafka.server.KafkaApis.handle(KafkaApis.scala:186)
 
 at kafka.server.KafkaRequestHandler.run(KafkaRequestHandler.scala:42)
 
 at java.lang.Thread.run(Thread.java:744)
 
 --
 Kushan Maskey
 
 
 
 
 -- 
 Regards,
 Vikas Agarwal
 91 – 9928301411
 
 InfoObjects, Inc. 
 Execution Matters
 http://www.infoobjects.com 
 2041 Mission College Boulevard, #280 
 Santa Clara, CA 95054
 +1 (408) 988-2000 Work
 +1 (408) 716-2726 Fax
 
 


-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.


Re: adding bolt to TridentTopology

2014-08-28 Thread Parth Brahmbhatt
You can not use storm core bolts to process trident tuples. You could use core 
spouts in trident topologies but not  bolts.
You will have to write your trident states that may do identical stuff as your 
bolt implementation in batch mode. You can refer to this doc to understand how 
you can write your own trident state 
https://storm.incubator.apache.org/documentation/Trident-tutorial.html

Hope this helps.

Thanks
Parth

On Aug 27, 2014, at 10:05 PM, Naga Vij nvbuc...@gmail.com wrote:

 Hello,
 
 How can I add a bolt to TridentTopology?
 
 I realize the use of function on the stream (kafka stream in my case) ...
 
 parsedStream.each(parsedStream.getOutputFields(), new SomeFunction(), new 
 Fields());
 
 But I want to do subsequent processing by using bolts after the function.
 
 How can I do that?  Tried looking into the API, but appears I need some help 
 from others who might have tried it already.
 
 Thanks in advance.
 
 Naga


-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.


Re: [DISCUSS] Apache Storm Release 0.9.3/0.10.0

2014-08-28 Thread Parth Brahmbhatt
I agree with the version bump and also with the strategy to have a beta 
release. 




Thanks

Parth
—
Sent from Mailbox

On Thu, Aug 28, 2014 at 2:59 PM, Derek Dagit der...@yahoo-inc.com wrote:

 I am supportive.
 I think it makes sense to move to 0.10.0 because of the significance of the 
 changes.
 -- 
 Derek
 On 8/28/14, 15:34, P.Taylor Goetz wrote:
 I’d like to gather community feedback for the next two releases of Apache 
 Storm.

 0.9.3-incubating will be our next release. Please indicate (by JIRA ticket 
 ID) which bug fixes and/or new features you would like to be considered for 
 inclusion in the next release. If there is not an existing for a particular 
 issue or feature, please consider adding one.

 For the next and subsequent releases, we will be using a slightly different 
 approach than what we did in the past. Instead of voting right away on a 
 build, we will make one or more “unofficial” release candidate builds 
 available prior to voting on an official release. This will give the Apache 
 Storm community more time to discuss, evaluate, identify and fix potential 
 issues before the official release. This should enable us to ensure the 
 final release is as bug free as possible.

 Apache Storm 0.10.0 (STORM-216)
 As some of you are aware, the engineering team at Yahoo! has done a lot of 
 work to bring security and multi-tenancy to Storm, and has contributed that 
 work back to the community. Over the past few months we have been in the 
 process of enhancing and syncing that work  with the master branch in a 
 separate branch labeled “security.” That work is now nearing completion, and 
 I would like us to consider merging it into master after the 0.9.3 release. 
 Since the security work includes a large number of changes and enhancements, 
 I propose we bump the version number to 0.10.0 for the first release to 
 include those features.

 More information about the security branch can be found in this pull request 
 [1], as well as the SECURITY.md file in the security branch [2]. I also 
 discussed it in a blog post [3] on the Hortonworks website. Please feel free 
 to direct any comments or questions about the security branch to the mailing 
 list.

 Similar to the process we’ll follow for 0.9.3, we plan to make several 
 unofficial “development” builds available for those who would like to help 
 with testing the new security features.


 -Taylor

 [1] https://github.com/apache/incubator-storm/pull/121
 [2] https://github.com/apache/incubator-storm/blob/security/SECURITY.md
 [3] http://hortonworks.com/blog/the-future-of-apache-storm/

-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.


Re: Storm of Maven issue?

2014-08-20 Thread Parth Brahmbhatt
Can you send you pom.xml? 

Thanks
Parth
On Aug 20, 2014, at 10:53 AM, Kevin Hill hill.ke...@gmail.com wrote:

 Hello all, I’m brand new to Storm (and Maven) so this might be a more basic 
 issue, but I’d love a bit of help. I’m trying to make my own project to house 
 some topos. My understanding is that the best way to do this is to use Maven 
 and include storm as a dependency.  The docs here: 
 https://storm.incubator.apache.org/documentation/Maven.html seem a bit out of 
 date, and I think missing some formatting for xml file code snippets, but 
 I’ve searched around and think they are now hosted at org.apache.storm:storm. 
 My problem is that Maven can’t seem to find them at any location. I get the 
 following error message whenever I try to compile with ‘mvn compile'
 
 
 
 [ERROR] Failed to execute goal on project rtspCluster: Could not resolve 
 dependencies for project fitnet:rtspCluster:jar:0.1-alpha: Could not find 
 artifact org.apache.storm:storm:jar:0.9.2-incubating - [Help 1]”
 
 
 
 I’ve also tried other storm repos I’ve found at storm:storm:0.9.0.1 etc, all 
 to the same effect. I’m guessing if I ask on Maven lists they will just say 
 ‘contact the repo maintainers’ but it seems like this must be something super 
 basic. What am I missing?
 
 
 
 -Kevin
 


-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.


Re: Reading config.priperties file

2014-08-20 Thread Parth Brahmbhatt
Are you packing the config file in the jar? Does the config file get loaded on 
the gateway , where you run storm command, or its suppose to be loaded as part 
of spout’s/bolt's prepare method? In the former case you need to ensure that 
your property file is part of your jar file. If you are using maven you can do 
so by adding the following to your build target:
resources
resource
directorysrc/resource/directory
/resource
/resources

and then in your code you can get a handle on the config file by 

SomeClass.class.getClassLoader().getResourceAsStream(“yourconfig.properties)


The other way would be to just read the config file at the gateway, read the 
properties and set the property value as instance variables in the appropriate 
bolt and spout object. Ensure that the instance variables are not marked as 
transient.

Thanks
Parth

If the config file is read and loaded at the gateway then are you storing 
On Aug 20, 2014, at 2:50 PM, Kushan Maskey 
kushan.mas...@mmillerassociates.com wrote:

 I pass the config file as an argument to the Topology.
 


-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.


Re: kafka.trident.ZkBrokerReader-brokers need refreshing

2014-08-13 Thread Parth Brahmbhatt
6 ms is an internal setting for zk refresh see 
https://github.com/apache/incubator-storm/blob/master/external/storm-kafka/src/jvm/storm/kafka/ZkHosts.java#L30.

Thanks
Parth

On Aug 13, 2014, at 9:44 AM, Sa Li sa.in.v...@gmail.com wrote:

 Hi, Siddharth
 
 I only used trident topology for kafka spout, since I thought I can easily to 
 add .each function to parse the stream. I do add two types of storm-kafka 
 packages in my pom, See this pom
 
  !-- Storm-Kafka compiled --
 dependency
 artifactIdstorm-kafka/artifactId
 groupIdorg.apache.storm/groupId
 version0.9.2-incubating/version
   !--
 scope*compile*/scope
   --
!-- exclude the zookeeper package from storm-Kafka --
 exclusions
exclusion
groupIdorg.apache.zookeeper/groupId
artifactIdzookeeper/artifactId
/exclusion
 /exclusions
 /dependency
 
  dependency
 groupIdstorm/groupId
 artifactIdstorm-kafka/artifactId
 version0.9.0-wip16a-scala292/version
 !-- exclude the zookeeper package from storm-Kafka 
 --
 exclusions
exclusion
groupIdorg.apache.zookeeper/groupId
artifactIdzookeeper/artifactId
/exclusion
 /exclusions
  /dependency
 
 When I firstly running the kafkaSpout, I got stuck for a while, after talking 
 to developers here back and forth, I realize the version conflict is really 
 an issue we need to pay attention, you must be make the zookeeper, storm, 
 kafka version consistent, otherwise you will have problem, or you need to 
 exclude it in pom.
 
 Thanks
 
 Alec
 
 
 On Wed, Aug 13, 2014 at 7:20 AM, siddharth ubale siddharth.ub...@gmail.com 
 wrote:
 hi ,
 
 Just curious, did u face any isue with using kafka Spout if u did not use 
 trident?
 Are u also able to implement the KafkaSpout packaged with Storm ? 
 I am asking cos i am unable to use the kafkaSpout(SpoutConfig) to read from 
 kafka topic. i am using kafka 8.1.1 and storm is 9.0.1 
 Can you lemme know about any issue u faced??
 
 i get no error while i submitting my program but only a never ending sequence 
 of the follwing:
 68582 [Thread-33-words] INFO  backtype.storm.daemon.executor - Processing 
 received message source: __system:-1, stream: __metrics_tick, id: {}, [60]
 68588 [Thread-33-words] INFO  backtype.storm.daemon.task - Emitting: words 
 __metrics [#TaskInfo 
 backtype.storm.metric.api.IMetricsConsumer$TaskInfo@a92bbf9 [#DataPoint 
 [__ack-count = {}] #DataPoint [__sendqueue = {write_pos=-1, read_pos=-1, 
 capacity=1024, population=0}] #DataPoint [__complete-latency = {}] 
 #DataPoint [__receive = {write_pos=3, read_pos=2, capacity=1024, 
 population=1}] #DataPoint [kafkaOffset = {totalLatestTime=0, 
 totalSpoutLag=0, totalLatestEmittedOffset=0}] #DataPoint [__transfer-count 
 = {}] #DataPoint [__fail-count = {}] #DataPoint [__emit-count = {}]]]
 68649 [Thread-33-words] INFO  storm.kafka.ZkCoordinator - Refreshing 
 partition manager connections
 68649 [Thread-34-words] INFO  storm.kafka.ZkCoordinator - Refreshing 
 partition manager connections
 68664 [Thread-33-words] INFO  storm.kafka.ZkCoordinator - Deleted partition 
 managers: []
 68664 [Thread-33-words] INFO  storm.kafka.ZkCoordinator - New partition 
 managers: []
 68664 [Thread-33-words] INFO  storm.kafka.ZkCoordinator - Finished refreshing
 68665 [Thread-34-words] INFO  storm.kafka.ZkCoordinator - Deleted partition 
 managers: []
 68665 [Thread-34-words] INFO  storm.kafka.ZkCoordinator - New partition 
 managers: []
 68665 [Thread-34-words] INFO  storm.kafka.ZkCoordinator - Finished refreshing
 
 Thanks,
 Siddharth
 
 
 
 On Wed, Aug 13, 2014 at 5:33 AM, Sa Li sa.in.v...@gmail.com wrote:
 Hi, All
 
 I am reading the messages from producer and print the time and 
 userhostaddress, but I am getting such warning once in a while:
 
 184.146.220.124
 1403070062
 24.79.224.172
 1403070063
 71.199.4.138
 2644780 [Thread-16-spout0] WARN  storm.artemis.kafka.KafkaUtils - No data 
 found in Kafka Partition partition_2
 1403070064
 172.4.221.83
 2647191 [Thread-16-spout0] INFO  storm.artemis.kafka.trident.ZkBrokerReader - 
 brokers need refreshing because 6ms have expired
 2647195 [Thread-16-spout0] INFO  storm.artemis.kafka.DynamicBrokersReader - 
 Read partition info from zookeeper: 
 GlobalPartitionInformation{partitionMap={0=10.100.70.128:9092, 
 1=10.100.70.128:9092, 2=10.100.70.128:9092, 3=10.100.70.128:9092, 
 4=10.100.70.128:9092}}
 2648569 [Thread-8-$spoutcoord-spout0] INFO  
 storm.artemis.kafka.trident.ZkBrokerReader - brokers need refreshing because 
 6ms have expired
 2648573 [Thread-8-$spoutcoord-spout0] INFO  
 storm.artemis.kafka.DynamicBrokersReader - Read partition info from 
 zookeeper: 

Re: kafka-spout running error

2014-08-05 Thread Parth Brahmbhatt
Can you let the topology run for 120 seconds or so? In my experience the kafka 
bolt/spout takes a lot of latency initially as it tries to read/write from 
zookeeper and initialize connections. On my mac it takes about 15 seconds 
before the spout is actually opened. 

Thanks
Parth
On Aug 5, 2014, at 1:11 PM, Sa Li sa.in.v...@gmail.com wrote:

 If I set the sleep time as 1000 milisec, I got such error:
 
 3067 [main] INFO  backtype.storm.testing - Deleting temporary path 
 /tmp/0f1851f1-9499-48a5-817e-41712921d054
 3163 [Thread-10-EventThread] INFO  
 com.netflix.curator.framework.state.ConnectionStateManager - State change: 
 SUSPENDED
 3163 [ConnectionStateManager-0] WARN  
 com.netflix.curator.framework.state.ConnectionStateManager - There are no 
 ConnectionStateListeners registered.
 3164 [Thread-10-EventThread] WARN  backtype.storm.cluster - Received event 
 :disconnected::none: with disconnected Zookeeper.
 3636 [Thread-10-SendThread(localhost:2000)] WARN  
 org.apache.zookeeper.ClientCnxn - Session 0x147a7c868ef000b for server null, 
 unexpected error, closing socket connection and attempting reconnect
 java.net.ConnectException: Connection refused
 at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) 
 ~[na:1.7.0_55]
 at 
 sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:739) 
 ~[na:1.7.0_55]
 at 
 org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1119) 
 ~[zookeeper-3.3.3.jar:3.3.3-1073969]
 4877 [Thread-10-SendThread(localhost:2000)] WARN  
 org.apache.zookeeper.ClientCnxn - Session 0x147a7c868ef000b for server null, 
 unexpected error, closing socket connection and attempting reconnect
 java.net.ConnectException: Connection refused
 at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method) 
 ~[na:1.7.0_55]
 at 
 sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:739) 
 ~[na:1.7.0_55]
 at 
 org.apache.zookeeper.ClientCnxn$SendThread.run(ClientCnxn.java:1119) 
 ~[zookeeper-3.3.3.jar:3.3.3-1073969]
 5566 [Thread-10-SendThread(localhost:2000)] WARN  
 org.apache.zookeeper.ClientCnxn - Session 0x147a7c868ef000b for server null, 
 unexpected error, closing socket connection and attempting reconnect
 java.net.ConnectException: Connection refused
 
 seems not even connected to zookeeper, any method to confirm to connection of 
 zookeeper?
 
 Thanks a lot
 
 Alec
 
 On Aug 5, 2014, at 12:58 PM, Sa Li sa.in.v...@gmail.com wrote:
 
 Thank you very much for your reply, Taylor. I tried to increase the sleep 
 time as 1 sec or 10 sec, however I got such error, it seems to be Async loop 
 error. Any idea about that?
 
 3053 [Thread-19-$spoutcoord-spout0] INFO  
 org.apache.curator.framework.imps.CuratorFrameworkImpl - Starting
 3058 [Thread-19-$spoutcoord-spout0] ERROR backtype.storm.util - Async loop 
 died!
 java.lang.NoSuchMethodError: 
 org.apache.zookeeper.ZooKeeper.init(Ljava/lang/String;ILorg/apache/zookeeper/Watcher;Z)V
 at 
 org.apache.curator.utils.DefaultZookeeperFactory.newZooKeeper(DefaultZookeeperFactory.java:29)
  ~[kafka-storm-bitmap-0.0.1-SNAPSHOT-jar-with-dependencies.jar:na]
 at 
 org.apache.curator.framework.imps.CuratorFrameworkImpl$2.newZooKeeper(CuratorFrameworkImpl.java:154)
  ~[kafka-storm-bitmap-0.0.1-SNAPSHOT-jar-with-dependencies.jar:na]
 at 
 org.apache.curator.HandleHolder$1.getZooKeeper(HandleHolder.java:94) 
 ~[kafka-storm-bitmap-0.0.1-SNAPSHOT-jar-with-dependencies.jar:na]
 at 
 org.apache.curator.HandleHolder.getZooKeeper(HandleHolder.java:55) 
 ~[kafka-storm-bitmap-0.0.1-SNAPSHOT-jar-with-dependencies.jar:na]
 at 
 org.apache.curator.ConnectionState.reset(ConnectionState.java:219) 
 ~[kafka-storm-bitmap-0.0.1-SNAPSHOT-jar-with-dependencies.jar:na]
 at 
 org.apache.curator.ConnectionState.start(ConnectionState.java:103) 
 ~[kafka-storm-bitmap-0.0.1-SNAPSHOT-jar-with-dependencies.jar:na]
 at 
 org.apache.curator.CuratorZookeeperClient.start(CuratorZookeeperClient.java:190)
  ~[kafka-storm-bitmap-0.0.1-SNAPSHOT-jar-with-dependencies.jar:na]
 at 
 org.apache.curator.framework.imps.CuratorFrameworkImpl.start(CuratorFrameworkImpl.java:256)
  ~[kafka-storm-bitmap-0.0.1-SNAPSHOT-jar-with-dependencies.jar:na]
 at 
 storm.artemis.kafka.DynamicBrokersReader.init(DynamicBrokersReader.java:36)
  ~[kafka-storm-bitmap-0.0.1-SNAPSHOT-jar-with-dependencies.jar:na]
 at 
 storm.artemis.kafka.trident.ZkBrokerReader.init(ZkBrokerReader.java:24) 
 ~[kafka-storm-bitmap-0.0.1-SNAPSHOT-jar-with-dependencies.jar:na]
 at 
 storm.artemis.kafka.KafkaUtils.makeBrokerReader(KafkaUtils.java:40) 
 ~[kafka-storm-bitmap-0.0.1-SNAPSHOT-jar-with-dependencies.jar:na]
 at 
 storm.artemis.kafka.trident.Coordinator.init(Coordinator.java:16) 
 ~[kafka-storm-bitmap-0.0.1-SNAPSHOT-jar-with-dependencies.jar:na]
 at 
 storm.artemis.kafka.trident.OpaqueTridentKafkaSpout.getCoordinator(OpaqueTridentKafkaSpout.java:29)
 

Re: Kafka Spout not reading from Topic

2014-07-30 Thread Parth Brahmbhatt
Hi,

Nothing really wrong with your code. Can you confirm that zookeeper is
indeed running on localhost:2181 and can you log into zookeeper CLI (zkCli)
and check that ls / returns brokers directory under which you should be
able to find your topic. This is the default directory that ZkHosts.java
looks for.

Thanks
Parth

On Tue, Jul 29, 2014 at 11:17 PM, Palak Shah spala...@gmail.com wrote:

 Hi,

 I am using the Kafka spout that in integrated in
 apache-storm-0.9.2-incubating release. I am able to submit the topology to
 my storm cluster, but it is not receiving any tuples from the Kafka topic.
 I know the topic (page_visits) has data because I can read it from the
 console.

 Here is my code for topology :

 public static void main(String[] args) throws AlreadyAliveException,
 InvalidTopologyException {
 BrokerHosts zkHost = new ZkHosts(localhost:2181);
 SpoutConfig spoutConfig = new SpoutConfig(
   zkHost, // list of Kafka brokers
   page_visits, // topic to read from
   /zkroot, // the root path in Zookeeper for the spout
 to store the consumer offsets
   discovery); // an id for this consumer for storing the
 consumer offsets in Zookeeper
 spoutConfig.forceFromStart = true;
 KafkaSpout kafkaSpout = new KafkaSpout(spoutConfig);

 TopologyBuilder builder = new TopologyBuilder();
 builder.setSpout(kafkaSpout, kafkaSpout);
 builder.setBolt(kafkaBolt, new
 PrinterBolt()).shuffleGrouping(kafkaSpout);

 Config conf = new Config();
 conf.setNumWorkers(4);
 conf.setDebug(true);

 StormSubmitter.submitTopology(args[0], conf,
 builder.createTopology());

 }

 }

 I am using apache-storm-0.9.2-incubating and kafka-0.9.2-0.8.1.1. Is this
 a versions compatibility issue? if so, which version should I use for this
 to work?

 Thanks in Advance,
 Palak Shah




-- 
Thanks
Parth

-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.


Re: Need help to use storm with mysql.

2014-07-18 Thread Parth Brahmbhatt
Hi Amjad,

Is there any reason you can not upgrade to hadoop 2.0? Hadoop 2.0 has made many 
improvements over 1.X versions and they are source compatible so any of your MR 
jobs will be unaffected as long as you recompile with 2.x.

The code we pointed at assumes that all the classes for hadoop 2.X are present 
in your class path. if you are not using maven or some other build system and 
would like to add jars manually you probably will have tough time resolving 
conflicts so I would advise against it.
If you still want to add jars manually my best guess would be to look under
YOUR_HADOO_INSTALLATION_DIR/libexec/share/hadoop/

Thanks
Parth
On Jul 18, 2014, at 10:56 AM, amjad khan amjadkhan987...@gmail.com wrote:

 Thanks for your reply taylor. I'm using hadoop1.0.2. Can u suggest me any 
 alternative to connect to hadoop.
 
 
 
 On Fri, Jul 18, 2014 at 8:45 AM, P. Taylor Goetz ptgo...@gmail.com wrote:
 What version of Hadoop are you using? Storm-hdfs requires Hadoop 2.x.
 
 - Taylor
 
 On Jul 18, 2014, at 6:07 AM, amjad khan amjadkhan987...@gmail.com wrote:
 
 Thanks for your help parth 
 
 When i trying to run the topology to write the data to hdfs it throws 
 exception Class Not Found: 
 org.apache.hadoop.client.hdfs.HDFSDataOutputStream$SyncFlags
 Can anyone tell me what are the jars needed to execute the code to write 
 data to hdfs. Please tell me all the required jars. 
 
 
 On Wed, Jul 16, 2014 at 10:46 AM, Parth Brahmbhatt 
 pbrahmbh...@hortonworks.com wrote:
 You can use 
 
 https://github.com/ptgoetz/storm-hdfs
 
 It supports writing to HDFS with both Storm bolts and trident states. 
 Thanks
 Parth
 
 On Jul 16, 2014, at 10:41 AM, amjad khan amjadkhan987...@gmail.com wrote:
 
 Can anyone provide the code for bolt to write its data to hdfs. Kindly tell 
 me the jar's required to run that bolt.
 
 
 On Mon, Jul 14, 2014 at 2:33 PM, Max Evers mcev...@gmail.com wrote:
 Can you expand on your use case? What is the query selecting on? Is the 
 column you are querying on indexed?  Do you really need to look at the 
 entire 20 gb every 20ms?
 
 On Jul 14, 2014 6:39 AM, amjad khan amjadkhan987...@gmail.com wrote:
 I made a storm topoogy where spout was fetching data from mysql using 
 select query. The select query was fired after every 30 msec but because 
 the size of the table is more than 20 GB the select query takes more than 
 10 sec to execute therefore this is not working. I need to know what are 
 the possible alternatives for this situation. Kindly reply as soon as 
 possible.
 
 Thanks, 
 
 
 
 CONFIDENTIALITY NOTICE
 NOTICE: This message is intended for the use of the individual or entity to 
 which it is addressed and may contain information that is confidential, 
 privileged and exempt from disclosure under applicable law. If the reader of 
 this message is not the intended recipient, you are hereby notified that any 
 printing, copying, dissemination, distribution, disclosure or forwarding of 
 this communication is strictly prohibited. If you have received this 
 communication in error, please contact the sender immediately and delete it 
 from your system. Thank You.
 
 
 


-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.