Yes, it would help others. :)
On Tue, Sep 9, 2014 at 6:40 PM, Stephen Hartzell <hartzell.step...@gmail.com > wrote: > I'm going to make a youtube tutorial to show how I got it to work. If > nothing else, it'll be for my reference. > > -Thanks, Stephen > > On Tue, Sep 9, 2014 at 9:07 AM, Vikas Agarwal <vi...@infoobjects.com> > wrote: > >> Good to know that it solved your problem. :) >> This is what open source community do. :) >> Even in this mail thread, I also learned and you contributed too. You >> shared the detailed steps to run the storm cluster. :) >> >> >> On Tue, Sep 9, 2014 at 6:28 PM, Stephen Hartzell < >> hartzell.step...@gmail.com> wrote: >> >>> HA HA HA HA!!!! Vikas, you crazy genius!!!!!!!!!!! It works!!!! It >>> works!!!!!!! :D :D :D :D. How much beer do I owe you? After hours and hours >>> and hours of work... it was a port issue... Thank you so much everyone for >>> your help and input! >>> >>> On Tue, Sep 9, 2014 at 8:27 AM, Vikas Agarwal <vi...@infoobjects.com> >>> wrote: >>> >>>> Did you open following ports in Ec2 security group: >>>> >>>> 2181 >>>> 3772 >>>> 3773 >>>> 6627 >>>> 6700 >>>> 6701 >>>> 6702 >>>> 6703 >>>> >>>> Some of the ports might not be required to be opened, but I guess try >>>> opening these ports for the same security group not for the world and start >>>> second instance within same security group. >>>> >>>> >>>> On Tue, Sep 9, 2014 at 5:46 PM, Stephen Hartzell < >>>> hartzell.step...@gmail.com> wrote: >>>> >>>>> My apologies, I run "bin/storm supervisor &" at the end. That was a >>>>> bad copy and paste. >>>>> >>>>> On Tue, Sep 9, 2014 at 8:10 AM, Vikas Agarwal <vi...@infoobjects.com> >>>>> wrote: >>>>> >>>>>> >>*I launch another basic EC2 CentOS machine and do the exact same >>>>>> commands except that I don't install zookeeper and I only run "bin/storm >>>>>> nimbus &" at the end*<< >>>>>> >>>>>> Why would you run nimbus on two machines? >>>>>> >>>>>> Yes, by default Hortonworks provides one supervisor node par machine, >>>>>> so I am trying to add new EC2 machine with supervisor role. And it worked >>>>>> upto some extent but still not fully functional. I am still testing its >>>>>> linear scalability. >>>>>> >>>>>> >>>>>> On Tue, Sep 9, 2014 at 5:15 PM, Stephen Hartzell < >>>>>> hartzell.step...@gmail.com> wrote: >>>>>> >>>>>>> Vikas, >>>>>>> >>>>>>> >>>>>>> I've tried to use the HortonWorks distribution, but that only >>>>>>> provides one supervisor and nimbus on one virtual machine. I'm excited >>>>>>> to >>>>>>> hear that you have storm working on AWS EC2 machines because that is >>>>>>> exactly what I am trying to do! Right now we're still in the development >>>>>>> stage, so all we are trying to do is to have one worker machine connect >>>>>>> to >>>>>>> one nimbus machine. So far we haven't got this work. >>>>>>> >>>>>>> Although it might be lengthy, let me go ahead and post the >>>>>>> commands I'm using to setup the nimbus machine. >>>>>>> >>>>>>> >>>>>>> *I launch a basic EC2 CentOS machine with ports 6627 and 8080 open >>>>>>> to TCP connections (it's public IP is 54.68.149.181)* >>>>>>> >>>>>>> sudo yum update >>>>>>> >>>>>>> sudo yum install libuuid-devel gcc gcc-c++ kernel-devel >>>>>>> >>>>>>> *# Install zeromq* >>>>>>> >>>>>>> wget http://download.zeromq.org/zeromq-2.1.7.tar.gz >>>>>>> >>>>>>> tar –zxvf zeromq-2.1.7.tar.gz >>>>>>> >>>>>>> cd zeromq-2.1.7.tar.gz >>>>>>> >>>>>>> ./configure >>>>>>> >>>>>>> make >>>>>>> >>>>>>> sudo make install >>>>>>> >>>>>>> cd ~ >>>>>>> >>>>>>> *# Install jzmq* >>>>>>> >>>>>>> sudo yum install git java-devel libtool >>>>>>> >>>>>>> export JAVA_HOME=/usr/lib/jvm/java-1.7.0-openjdk-1.7.0.x.x86_64 *# >>>>>>> JDK required for jzmq in the configuration stage* >>>>>>> >>>>>>> cd $JAVA_HOME >>>>>>> >>>>>>> sudo ln –s include include *# >>>>>>> JDK include directory has headers required for jzmq* >>>>>>> >>>>>>> cd ~ >>>>>>> >>>>>>> git clone https://github.com/nathanmarz/jzmq.git >>>>>>> >>>>>>> cd jzmq/src/ >>>>>>> >>>>>>> CLASSPATH=.:./.:$CLASSPATH >>>>>>> >>>>>>> touch classdist_noinst.stamp >>>>>>> >>>>>>> javac -d . org/zeromq/ZMQ.java org/zeromq/ZMQException.java >>>>>>> org/zeromq/ZMQQueue.java org/zeromq/ZMQForwarder.java >>>>>>> org/zeromq/ZMQStreamer.java >>>>>>> >>>>>>> cd ~/jzmq >>>>>>> >>>>>>> ./autogen.sh >>>>>>> >>>>>>> ./configure >>>>>>> >>>>>>> make >>>>>>> >>>>>>> sudo make install >>>>>>> >>>>>>> cd ~ >>>>>>> >>>>>>> *# Download zookeeper* >>>>>>> >>>>>>> wget >>>>>>> http://mirror.cc.columbia.edu/pub/software/apache/zookeeper/zookeeper-3.4.6/zookeeper-3.4.6.tar.gz >>>>>>> >>>>>>> tar -zxvf zookeeper-3.4.6.tar.gz >>>>>>> >>>>>>> cd zookeeper-3.4.6 >>>>>>> >>>>>>> vi conf/zoo.cfg >>>>>>> >>>>>>> >>>>>>> tickTime=2000 >>>>>>> >>>>>>> dataDir=/tmp/zookeeper >>>>>>> >>>>>>> clientPort=2181 >>>>>>> >>>>>>> >>>>>>> mkdir /tmp/zookeeper >>>>>>> bin/zkServer.sh start >>>>>>> bin/zkCli.sh -server 54.68.149.181:2181 >>>>>>> >>>>>>> >>>>>>> *# Download storm and modify configuration file (changes over >>>>>>> default.yaml shown in bold)* >>>>>>> >>>>>>> https://www.dropbox.com/s/fl4kr7w0oc8ihdw/storm-0.8.2.zipunzip >>>>>>> storm-0.8.2.zip >>>>>>> <https://www.dropbox.com/s/fl4kr7w0oc8ihdw/storm-0.8.2.zipunzipstorm-0.8.2.zip> >>>>>>> >>>>>>> cd storm-0.8.2.zip >>>>>>> >>>>>>> vi conf/storm.yaml >>>>>>> >>>>>>> >>>>>>> # Licensed to the Apache Software Foundation (ASF) under one >>>>>>>> >>>>>>>> # or more contributor license agreements. See the NOTICE file >>>>>>>> >>>>>>>> # distributed with this work for additional information >>>>>>>> >>>>>>>> # regarding copyright ownership. The ASF licenses this file >>>>>>>> >>>>>>>> # to you under the Apache License, Version 2.0 (the >>>>>>>> >>>>>>>> # "License"); you may not use this file except in compliance >>>>>>>> >>>>>>>> # with the License. You may obtain a copy of the License at >>>>>>>> >>>>>>>> # >>>>>>>> >>>>>>>> # http:# www.apache.org/licenses/LICENSE-2.0 >>>>>>>> >>>>>>>> # >>>>>>>> >>>>>>>> # Unless required by applicable law or agreed to in writing, >>>>>>>>> software >>>>>>>> >>>>>>>> # distributed under the License is distributed on an "AS IS" BASIS, >>>>>>>> >>>>>>>> # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or >>>>>>>>> implied. >>>>>>>> >>>>>>>> # See the License for the specific language governing permissions >>>>>>>>> and >>>>>>>> >>>>>>>> # limitations under the License. >>>>>>>> >>>>>>>> >>>>>>>>> >>>>>>>>> ########### These all have default values as shown >>>>>>>> >>>>>>>> ########### Additional configuration goes into storm.yaml >>>>>>>> >>>>>>>> >>>>>>>>> java.library.path: "/usr/local/lib:/opt/local/lib:/usr/lib: >>>>>>>>> */home/ec2-user*" >>>>>>>> >>>>>>>> >>>>>>>>> ### storm.* configs are general configurations >>>>>>>> >>>>>>>> # the local dir is where jars are kept >>>>>>>> >>>>>>>> storm.local.dir: "storm-local" >>>>>>>> >>>>>>>> storm.zookeeper.servers: >>>>>>>> >>>>>>>> - "*54.68.149.181*" >>>>>>>> >>>>>>>> storm.zookeeper.port: 2181 >>>>>>>> >>>>>>>> storm.zookeeper.root: "/storm" >>>>>>>> >>>>>>>> storm.zookeeper.session.timeout: 20000 >>>>>>>> >>>>>>>> storm.zookeeper.connection.timeout: 15000 >>>>>>>> >>>>>>>> storm.zookeeper.retry.times: 5 >>>>>>>> >>>>>>>> storm.zookeeper.retry.interval: 1000 >>>>>>>> >>>>>>>> storm.zookeeper.retry.intervalceiling.millis: 30000 >>>>>>>> >>>>>>>> storm.cluster.mode: "distributed" # can be distributed or local >>>>>>>> >>>>>>>> storm.local.mode.zmq: false >>>>>>>> >>>>>>>> storm.thrift.transport: >>>>>>>>> "backtype.storm.security.auth.SimpleTransportPlugin" >>>>>>>> >>>>>>>> storm.messaging.transport: "backtype.storm.messaging.zmq" >>>>>>>> >>>>>>>> >>>>>>>>> ### nimbus.* configs are for the master >>>>>>>> >>>>>>>> nimbus.host: "*54.68.149.181*" >>>>>>>> >>>>>>>> nimbus.thrift.port: 6627 >>>>>>>> >>>>>>>> nimbus.childopts: "-Xmx1024m" >>>>>>>> >>>>>>>> nimbus.task.timeout.secs: 30 >>>>>>>> >>>>>>>> nimbus.supervisor.timeout.secs: 60 >>>>>>>> >>>>>>>> nimbus.monitor.freq.secs: 10 >>>>>>>> >>>>>>>> nimbus.cleanup.inbox.freq.secs: 600 >>>>>>>> >>>>>>>> nimbus.inbox.jar.expiration.secs: 3600 >>>>>>>> >>>>>>>> nimbus.task.launch.secs: 120 >>>>>>>> >>>>>>>> nimbus.reassign: true >>>>>>>> >>>>>>>> nimbus.file.copy.expiration.secs: 600 >>>>>>>> >>>>>>>> nimbus.topology.validator: >>>>>>>>> "backtype.storm.nimbus.DefaultTopologyValidator" >>>>>>>> >>>>>>>> >>>>>>>>> ### ui.* configs are for the master >>>>>>>> >>>>>>>> ui.port: 8080 >>>>>>>> >>>>>>>> ui.childopts: "-Xmx768m" >>>>>>>> >>>>>>>> >>>>>>>>> logviewer.port: 8000 >>>>>>>> >>>>>>>> logviewer.childopts: "-Xmx128m" >>>>>>>> >>>>>>>> logviewer.appender.name: "A1" >>>>>>>> >>>>>>>> >>>>>>>>> >>>>>>>>> drpc.port: 3772 >>>>>>>> >>>>>>>> drpc.worker.threads: 64 >>>>>>>> >>>>>>>> drpc.queue.size: 128 >>>>>>>> >>>>>>>> drpc.invocations.port: 3773 >>>>>>>> >>>>>>>> drpc.request.timeout.secs: 600 >>>>>>>> >>>>>>>> drpc.childopts: "-Xmx768m" >>>>>>>> >>>>>>>> >>>>>>>>> transactional.zookeeper.root: "/transactional" >>>>>>>> >>>>>>>> transactional.zookeeper.servers: null >>>>>>>> >>>>>>>> transactional.zookeeper.port: null >>>>>>>> >>>>>>>> >>>>>>>>> ### supervisor.* configs are for node supervisors >>>>>>>> >>>>>>>> # Define the amount of workers that can be run on this machine. >>>>>>>>> Each worker is assigned a port to use for communication >>>>>>>> >>>>>>>> supervisor.slots.ports: >>>>>>>> >>>>>>>> - 6700 >>>>>>>> >>>>>>>> - 6701 >>>>>>>> >>>>>>>> - 6702 >>>>>>>> >>>>>>>> - 6703 >>>>>>>> >>>>>>>> supervisor.childopts: "-Xmx256m" >>>>>>>> >>>>>>>> #how long supervisor will wait to ensure that a worker process is >>>>>>>>> started >>>>>>>> >>>>>>>> supervisor.worker.start.timeout.secs: 120 >>>>>>>> >>>>>>>> #how long between heartbeats until supervisor considers that worker >>>>>>>>> dead and tries to restart it >>>>>>>> >>>>>>>> supervisor.worker.timeout.secs: 30 >>>>>>>> >>>>>>>> #how frequently the supervisor checks on the status of the >>>>>>>>> processes it's monitoring and restarts if necessary >>>>>>>> >>>>>>>> supervisor.monitor.frequency.secs: 3 >>>>>>>> >>>>>>>> #how frequently the supervisor heartbeats to the cluster state (for >>>>>>>>> nimbus) >>>>>>>> >>>>>>>> supervisor.heartbeat.frequency.secs: 5 >>>>>>>> >>>>>>>> supervisor.enable: true >>>>>>>> >>>>>>>> >>>>>>>>> ### worker.* configs are for task workers >>>>>>>> >>>>>>>> worker.childopts: "-Xmx768m" >>>>>>>> >>>>>>>> worker.heartbeat.frequency.secs: 1 >>>>>>>> >>>>>>>> >>>>>>>>> task.heartbeat.frequency.secs: 3 >>>>>>>> >>>>>>>> task.refresh.poll.secs: 10 >>>>>>>> >>>>>>>> >>>>>>>>> zmq.threads: 1 >>>>>>>> >>>>>>>> zmq.linger.millis: 5000 >>>>>>>> >>>>>>>> zmq.hwm: 0 >>>>>>>> >>>>>>>> >>>>>>>>> >>>>>>>>> storm.messaging.netty.server_worker_threads: 1 >>>>>>>> >>>>>>>> storm.messaging.netty.client_worker_threads: 1 >>>>>>>> >>>>>>>> storm.messaging.netty.buffer_size: 5242880 #5MB buffer >>>>>>>> >>>>>>>> storm.messaging.netty.max_retries: 30 >>>>>>>> >>>>>>>> storm.messaging.netty.max_wait_ms: 1000 >>>>>>>> >>>>>>>> storm.messaging.netty.min_wait_ms: 100 >>>>>>>> >>>>>>>> >>>>>>>>> ### topology.* configs are for specific executing storms >>>>>>>> >>>>>>>> topology.enable.message.timeouts: true >>>>>>>> >>>>>>>> topology.debug: false >>>>>>>> >>>>>>>> topology.optimize: true >>>>>>>> >>>>>>>> topology.workers: 1 >>>>>>>> >>>>>>>> topology.acker.executors: null >>>>>>>> >>>>>>>> topology.tasks: null >>>>>>>> >>>>>>>> # maximum amount of time a message has to complete before it's >>>>>>>>> considered failed >>>>>>>> >>>>>>>> topology.message.timeout.secs: 30 >>>>>>>> >>>>>>>> topology.skip.missing.kryo.registrations: false >>>>>>>> >>>>>>>> topology.max.task.parallelism: null >>>>>>>> >>>>>>>> topology.max.spout.pending: null >>>>>>>> >>>>>>>> topology.state.synchronization.timeout.secs: 60 >>>>>>>> >>>>>>>> topology.stats.sample.rate: 0.05 >>>>>>>> >>>>>>>> topology.builtin.metrics.bucket.size.secs: 60 >>>>>>>> >>>>>>>> topology.fall.back.on.java.serialization: true >>>>>>>> >>>>>>>> topology.worker.childopts: null >>>>>>>> >>>>>>>> topology.executor.receive.buffer.size: 1024 #batched >>>>>>>> >>>>>>>> topology.executor.send.buffer.size: 1024 #individual messages >>>>>>>> >>>>>>>> topology.receiver.buffer.size: 8 # setting it too high causes a lot >>>>>>>>> of problems (heartbeat thread gets starved, throughput plummets) >>>>>>>> >>>>>>>> topology.transfer.buffer.size: 1024 # batched >>>>>>>> >>>>>>>> topology.tick.tuple.freq.secs: null >>>>>>>> >>>>>>>> topology.worker.shared.thread.pool.size: 4 >>>>>>>> >>>>>>>> topology.disruptor.wait.strategy: >>>>>>>>> "com.lmax.disruptor.BlockingWaitStrategy" >>>>>>>> >>>>>>>> topology.spout.wait.strategy: >>>>>>>>> "backtype.storm.spout.SleepSpoutWaitStrategy" >>>>>>>> >>>>>>>> topology.sleep.spout.wait.strategy.time.ms: 1 >>>>>>>> >>>>>>>> topology.error.throttle.interval.secs: 10 >>>>>>>> >>>>>>>> topology.max.error.report.per.interval: 5 >>>>>>>> >>>>>>>> topology.kryo.factory: >>>>>>>>> "backtype.storm.serialization.DefaultKryoFactory" >>>>>>>> >>>>>>>> topology.tuple.serializer: >>>>>>>>> "backtype.storm.serialization.types.ListDelegateSerializer" >>>>>>>> >>>>>>>> topology.trident.batch.emit.interval.millis: 500 >>>>>>>> >>>>>>>> >>>>>>>>> dev.zookeeper.path: "/tmp/dev-storm-zookeeper" >>>>>>>> >>>>>>>> >>>>>>> >>>>>>> bin/storm nimbus & >>>>>>> >>>>>>> bin/storm supervisor & >>>>>>> >>>>>>> bin/storm ui & >>>>>>> >>>>>>> *I launch another basic EC2 CentOS machine and do the exact same >>>>>>> commands except that I don't install zookeeper and I only run "bin/storm >>>>>>> nimbus &" at the end* >>>>>>> >>>>>>> Any thoughts would be greatly appreciated. Thanks so much for all of >>>>>>> your help. I'm sure someone else has done this before! >>>>>>> >>>>>>> On Mon, Sep 8, 2014 at 11:14 PM, Vikas Agarwal < >>>>>>> vi...@infoobjects.com> wrote: >>>>>>> >>>>>>>> Although, implementing the Storm cluster manually would be really >>>>>>>> nice learning, I would suggest using HortonWorks distribution which >>>>>>>> comes >>>>>>>> with Storm as OOTB solution and you can configure everything from >>>>>>>> Ambari >>>>>>>> UI. We are using Storm on Amazon EC2 machine, though it is right now in >>>>>>>> beta stage. We are going to move to production in coming 2-3 months. >>>>>>>> >>>>>>>> >>>>>>>> On Tue, Sep 9, 2014 at 5:53 AM, Stephen Hartzell < >>>>>>>> hartzell.step...@gmail.com> wrote: >>>>>>>> >>>>>>>>> All, >>>>>>>>> >>>>>>>>> I implemented the suggestions given by Parh and Harsha. I am now >>>>>>>>> using the default.yaml but I changed the storm.zookeeper.servers to >>>>>>>>> the >>>>>>>>> nimbus machine's ip address: 54.68.149.181. I also changed the >>>>>>>>> nimbus.host >>>>>>>>> to 54.68.149.181. I also opened up port 6627. Now, the UI web page >>>>>>>>> gives >>>>>>>>> the following error: org.apache.thrift7.transport.TTransportException: >>>>>>>>> java.net.ConnectException: Connection refused >>>>>>>>> >>>>>>>>> You should be able to see the error it gives by going to the web >>>>>>>>> page yourself at: http://54.68.149.181:8080. I am only using this >>>>>>>>> account to test and see if I can even get storm to work, so these >>>>>>>>> machines >>>>>>>>> are only for testing. Perhaps someone could tell me what the >>>>>>>>> storm.yaml >>>>>>>>> file should look like for this setup? >>>>>>>>> >>>>>>>>> -Thanks, Stephne >>>>>>>>> >>>>>>>>> On Mon, Sep 8, 2014 at 7:41 PM, Stephen Hartzell < >>>>>>>>> hartzell.step...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> I'm getting kind of confused by the storm.yaml file. Should I be >>>>>>>>>> using the default.yaml and just modify the zookeeper and nimbus ip, >>>>>>>>>> or >>>>>>>>>> should I use a bran new storm.yaml? >>>>>>>>>> >>>>>>>>>> My nimbus machine has the ip address: 54.68.149.181. My >>>>>>>>>> zookeeper is on the nimbus machine. what should the storm.yaml look >>>>>>>>>> like on >>>>>>>>>> my worker and nimbus machine? Will the storm.yaml be the same on my >>>>>>>>>> worker >>>>>>>>>> and nimbus machine? I am not trying to do anything fancy, I am just >>>>>>>>>> trying >>>>>>>>>> to get a very basic cluster up and running. >>>>>>>>>> >>>>>>>>>> -Thanks, Stephen >>>>>>>>>> >>>>>>>>>> On Mon, Sep 8, 2014 at 7:00 PM, Stephen Hartzell < >>>>>>>>>> hartzell.step...@gmail.com> wrote: >>>>>>>>>> >>>>>>>>>>> All Thanks so much for your help. I cannot tell you how much I >>>>>>>>>>> appreciate it. I'm going to try out your suggestions and keep >>>>>>>>>>> banging my >>>>>>>>>>> head again the wall : D. I've spent an enormous amount of time >>>>>>>>>>> trying to >>>>>>>>>>> get this to work. I'll let you know what happens after I try to >>>>>>>>>>> implement >>>>>>>>>>> your suggestions. It would be really cool if someone had a tutorial >>>>>>>>>>> that >>>>>>>>>>> detailed this part. (I'll make it myself if I ever get this to >>>>>>>>>>> work!) It >>>>>>>>>>> seems like trying to get a two-machine cluster setup on AWS would >>>>>>>>>>> be a VERY >>>>>>>>>>> common use-case. I've read and watched everything I can on the >>>>>>>>>>> topic and >>>>>>>>>>> nothing got it working for me! >>>>>>>>>>> >>>>>>>>>>> On Mon, Sep 8, 2014 at 6:54 PM, Parth Brahmbhatt < >>>>>>>>>>> pbrahmbh...@hortonworks.com> wrote: >>>>>>>>>>> >>>>>>>>>>>> The worker connects to the thrift port and not the ui port. You >>>>>>>>>>>> need to open port 6627 or whatever is the value being set in >>>>>>>>>>>> storm.yaml >>>>>>>>>>>> using property “nimbus.thrift.port”. >>>>>>>>>>>> >>>>>>>>>>>> Based on the configuration that you have pointed so far it >>>>>>>>>>>> seems your nimbus host has nimbus,ui,supervisor working because you >>>>>>>>>>>> actually have zookeeper running locally on that host. As Harsha >>>>>>>>>>>> pointed out >>>>>>>>>>>> you need to change it to a value that is the public ip instead of >>>>>>>>>>>> loopback >>>>>>>>>>>> interface. >>>>>>>>>>>> >>>>>>>>>>>> Thanks >>>>>>>>>>>> Parth >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Sep 8, 2014, at 3:42 PM, Harsha <st...@harsha.io> wrote: >>>>>>>>>>>> >>>>>>>>>>>> storm.zookeeper.servers: >>>>>>>>>>>> - "127.0.0.1" >>>>>>>>>>>> nimbus.host: "127.0.0.1" ( *127.0.0.1 causes to bind a >>>>>>>>>>>> loopback interface , instead either use your public ip or 0.0.0.0* >>>>>>>>>>>> ) >>>>>>>>>>>> storm.local.dir: /tmp/storm ( I* recommend this to move to a >>>>>>>>>>>> different folder probably /home/storm, /tmp/storm will get deleted >>>>>>>>>>>> if your >>>>>>>>>>>> machine is restarted)* >>>>>>>>>>>> >>>>>>>>>>>> make sure you zookeeper is also listening in 0.0.0.0 or public >>>>>>>>>>>> ip not 127.0.0.1. >>>>>>>>>>>> >>>>>>>>>>>> "No, I cannot ping my host which has a public ip address of >>>>>>>>>>>> 54.68.149.181" >>>>>>>>>>>> you are not able to reach this ip form worker node but able to >>>>>>>>>>>> access the UI using it? >>>>>>>>>>>> -Harsha >>>>>>>>>>>> >>>>>>>>>>>> On Mon, Sep 8, 2014, at 03:34 PM, Stephen Hartzell wrote: >>>>>>>>>>>> >>>>>>>>>>>> Harsha, >>>>>>>>>>>> >>>>>>>>>>>> The storm.yaml on the host machine looks like this: >>>>>>>>>>>> >>>>>>>>>>>> storm.zookeeper.servers: >>>>>>>>>>>> - "127.0.0.1" >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> nimbus.host: "127.0.0.1" >>>>>>>>>>>> >>>>>>>>>>>> storm.local.dir: /tmp/storm >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> The storm.yaml on the worker machine looks like this: >>>>>>>>>>>> >>>>>>>>>>>> storm.zookeeper.servers: >>>>>>>>>>>> - "54.68.149.181" >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> nimbus.host: "54.68.149.181" >>>>>>>>>>>> >>>>>>>>>>>> storm.local.dir: /tmp/storm >>>>>>>>>>>> >>>>>>>>>>>> No, I cannot ping my host which has a public ip address of >>>>>>>>>>>> 54.68.149.181 although I can connect to the UI web page when it is >>>>>>>>>>>> hosted. >>>>>>>>>>>> I don't know how I would go about connecting to zookeeper on the >>>>>>>>>>>> nimbus >>>>>>>>>>>> host. >>>>>>>>>>>> -Thanks, Stephen >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Mon, Sep 8, 2014 at 6:28 PM, Harsha <st...@harsha.io> wrote: >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> There aren't any errors in worker machine supervisor logs. Are >>>>>>>>>>>> you using the same storm.yaml for both the machines and also are >>>>>>>>>>>> you able >>>>>>>>>>>> to ping your nimbus host or connect to zookeeper on nimbus host. >>>>>>>>>>>> -Harsha >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Mon, Sep 8, 2014, at 03:24 PM, Stephen Hartzell wrote: >>>>>>>>>>>> >>>>>>>>>>>> Harsha, >>>>>>>>>>>> >>>>>>>>>>>> Thanks so much for getting back with me. I will check the >>>>>>>>>>>> logs, but I don't seem to get any error messages. I have a nimbus >>>>>>>>>>>> AWS >>>>>>>>>>>> machine with zookeeper on it and a worker AWS machine. >>>>>>>>>>>> >>>>>>>>>>>> On the nimbus machine I start the zookeeper and then I run: >>>>>>>>>>>> >>>>>>>>>>>> bin/storm nimbus & >>>>>>>>>>>> bin/storm supervisor & >>>>>>>>>>>> bin/storm ui >>>>>>>>>>>> >>>>>>>>>>>> On the worker machine I run: >>>>>>>>>>>> bin/storm supervisor >>>>>>>>>>>> >>>>>>>>>>>> When I go to the UI page, I only see 1 supervisor (the one on >>>>>>>>>>>> the nimbus machine). So apparently, the worker machine isn't >>>>>>>>>>>> "registering" >>>>>>>>>>>> with the nimbus machine. >>>>>>>>>>>> >>>>>>>>>>>> On Mon, Sep 8, 2014 at 6:16 PM, Harsha <st...@harsha.io> wrote: >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> Hi Stephen, >>>>>>>>>>>> What are the issues you are seeing. >>>>>>>>>>>> "How do worker machines "know" how to connect to nimbus? Is it >>>>>>>>>>>> in the storm configuration file" >>>>>>>>>>>> >>>>>>>>>>>> Yes. make sure you the supervisor(worker) , nimbus nodes are >>>>>>>>>>>> able to connect to your zookeeper cluster. >>>>>>>>>>>> Check your logs under storm_inst/logs/ for any errors when you >>>>>>>>>>>> try to start nimbus or supervisors. >>>>>>>>>>>> If you are installing it manually try following these steps if >>>>>>>>>>>> you are not already done. >>>>>>>>>>>> >>>>>>>>>>>> http://www.michael-noll.com/tutorials/running-multi-node-storm-cluster/ >>>>>>>>>>>> -Harsha >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Mon, Sep 8, 2014, at 03:01 PM, Stephen Hartzell wrote: >>>>>>>>>>>> >>>>>>>>>>>> All, >>>>>>>>>>>> >>>>>>>>>>>> I would greatly appreciate any help that anyone would afford. >>>>>>>>>>>> I've been trying to setup a storm cluster on AWS for a few weeks >>>>>>>>>>>> now on >>>>>>>>>>>> centOS EC2 machines. So far, I haven't been able to get a cluster >>>>>>>>>>>> built. I >>>>>>>>>>>> can get a supervisor and nimbus to run on a single machine, but I >>>>>>>>>>>> can't >>>>>>>>>>>> figure out how to get another worker to connect to nimbus. How do >>>>>>>>>>>> worker >>>>>>>>>>>> machines "know" how to connect to nimbus? Is it in the storm >>>>>>>>>>>> configuration >>>>>>>>>>>> file? I've gone through many tutorials and the official >>>>>>>>>>>> documentation, but >>>>>>>>>>>> this point doesn't seem to be covered anywhere in sufficient >>>>>>>>>>>> detail for a >>>>>>>>>>>> new guy like me. >>>>>>>>>>>> >>>>>>>>>>>> Some of you may be tempted to point me toward storm-deploy, >>>>>>>>>>>> but I spent four days trying to get that to work until I gave up. >>>>>>>>>>>> I'm >>>>>>>>>>>> having Issue #58 on github. Following the instructions exactly and >>>>>>>>>>>> other >>>>>>>>>>>> tutorials on a bran new AWS machine fails. So I gave up on >>>>>>>>>>>> storm-deploy and >>>>>>>>>>>> decided to try and setup a cluster manually. Thanks in advance to >>>>>>>>>>>> anyone >>>>>>>>>>>> willing to offer me any inputs you can! >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> CONFIDENTIALITY NOTICE >>>>>>>>>>>> NOTICE: This message is intended for the use of the individual >>>>>>>>>>>> or entity to which it is addressed and may contain information >>>>>>>>>>>> that is >>>>>>>>>>>> confidential, privileged and exempt from disclosure under >>>>>>>>>>>> applicable law. >>>>>>>>>>>> If the reader of this message is not the intended recipient, you >>>>>>>>>>>> are hereby >>>>>>>>>>>> notified that any printing, copying, dissemination, distribution, >>>>>>>>>>>> disclosure or forwarding of this communication is strictly >>>>>>>>>>>> prohibited. If >>>>>>>>>>>> you have received this communication in error, please contact the >>>>>>>>>>>> sender >>>>>>>>>>>> immediately and delete it from your system. Thank You. >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> Regards, >>>>>>>> Vikas Agarwal >>>>>>>> 91 – 9928301411 >>>>>>>> >>>>>>>> InfoObjects, Inc. >>>>>>>> Execution Matters >>>>>>>> http://www.infoobjects.com >>>>>>>> 2041 Mission College Boulevard, #280 >>>>>>>> Santa Clara, CA 95054 >>>>>>>> +1 (408) 988-2000 Work >>>>>>>> +1 (408) 716-2726 Fax >>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> Regards, >>>>>> Vikas Agarwal >>>>>> 91 – 9928301411 >>>>>> >>>>>> InfoObjects, Inc. >>>>>> Execution Matters >>>>>> http://www.infoobjects.com >>>>>> 2041 Mission College Boulevard, #280 >>>>>> Santa Clara, CA 95054 >>>>>> +1 (408) 988-2000 Work >>>>>> +1 (408) 716-2726 Fax >>>>>> >>>>>> >>>>> >>>> >>>> >>>> -- >>>> Regards, >>>> Vikas Agarwal >>>> 91 – 9928301411 >>>> >>>> InfoObjects, Inc. >>>> Execution Matters >>>> http://www.infoobjects.com >>>> 2041 Mission College Boulevard, #280 >>>> Santa Clara, CA 95054 >>>> +1 (408) 988-2000 Work >>>> +1 (408) 716-2726 Fax >>>> >>>> >>> >> >> >> -- >> Regards, >> Vikas Agarwal >> 91 – 9928301411 >> >> InfoObjects, Inc. >> Execution Matters >> http://www.infoobjects.com >> 2041 Mission College Boulevard, #280 >> Santa Clara, CA 95054 >> +1 (408) 988-2000 Work >> +1 (408) 716-2726 Fax >> >> > -- Regards, Vikas Agarwal 91 – 9928301411 InfoObjects, Inc. Execution Matters http://www.infoobjects.com 2041 Mission College Boulevard, #280 Santa Clara, CA 95054 +1 (408) 988-2000 Work +1 (408) 716-2726 Fax