Hi Stig, Thank You for your help. I've run things again, and do see the numerous 'Emitted' statements and associated values now. I went back to my previous runs, and it looks like I had been getting the following messages at that point instead:
59817 [SLOT_1027] INFO o.a.s.d.s.Slot - STATE WAITING_FOR_WORKER_START msInState: 5 topo:test-1-1556727050 worker:772729c0-7c69-4905-9fee-a4ad4d7d5e50 -> RUNNING msInState: 0 topo:test-1-1556727050 worker:772729c0-7c69-4905-9fee-a4ad4d7d5e50 59817 [SLOT_1027] WARN o.a.s.d.s.Slot - SLOT 1027: Assignment Changed from LocalAssignment(topology_id:test-1-1556727050, executors:[ExecutorInfo(task_start:8, task_end:8), ExecutorInfo(task_start:12, task_end:12), ExecutorInfo(task_start:2, task_end:2), ExecutorInfo(task_start:6, task_end:6), ExecutorInfo(task_start:16, task_end:16), ExecutorInfo(task_start:10, task_end:10), ExecutorInfo(task_start:14, task_end:14), ExecutorInfo(task_start:4, task_end:4), ExecutorInfo(task_start:7, task_end:7), ExecutorInfo(task_start:3, task_end:3), ExecutorInfo(task_start:1, task_end:1), ExecutorInfo(task_start:9, task_end:9), ExecutorInfo(task_start:11, task_end:11), ExecutorInfo(task_start:13, task_end:13), ExecutorInfo(task_start:5, task_end:5), ExecutorInfo(task_start:15, task_end:15)], resources:WorkerResources(mem_on_heap:0.0, mem_off_heap:0.0, cpu:0.0), owner:david) to null 59817 [SLOT_1027] INFO o.a.s.ProcessSimulator - Begin killing process 772729c0-7c69-4905-9fee-a4ad4d7d5e50 59817 [SLOT_1027] INFO o.a.s.d.worker - Shutting down worker test-1-1556727050 8487bae7-a5b0-43f0-96eb-f198947e95e7 1027 I had been piping the output to a file through a '> output.file' at the end of the command line; I assume this was causing some kind of problem with starting or managing the local tasks. I had done the first runs without the pipe, but apparently missed the output. I appreciate your help, and can now move on to some more involved usage. Thanks again... Dave H. On Wed, May 1, 2019 at 5:14 PM Stig Rohde Døssing <[email protected]> wrote: > Hi David, > > I wouldn't worry about that log. It's not an error as far as I can tell. > The log is at INFO level as well. > > Storm-starter works for me. I get the same logs as you, but after "4987 > [main] INFO o.a.s.d.nimbus - Activating test: test-1-1556744624", I get a > bunch more log from Storm running the topology. Could you try with the > Storm 2.0.0 RC ( > https://dist.apache.org/repos/dist/dev/storm/apache-storm-2.0.0-rc7/)? > I'd be curious to know if this is also a problem there. > > Also just to be clear, are you getting nothing past the line " 48258 > [main] INFO o.a.s.d.nimbus - Activating test: test-1-1556727673"? > > Den ons. 1. maj 2019 kl. 22.43 skrev David Hutira <[email protected]>: > >> Hello, >> >> I'm new to Apache Storm, and have run into some trouble running the >> storm-starter ExclamationTopology example. I'm running on a MacBookPro with >> Mojave, I've downloaded and installed the Storm binary tar, and generated >> the storm-starter uber-jar with no (obvious) errors. When I run locally >> with the following command: >> >> storm jar storm-starter-1.2.2.jar >> org.apache.storm.starter.ExclamationTopology >> >> >> I get alot of output; bracketing the Local processing in main... >> >> >> *else* { >> >> >> LocalCluster cluster = *new* LocalCluster(); >> >> System.*out*.println("Starting Topology......"); >> >> cluster.submitTopology("test", conf, builder.createTopology()); >> >> System.*out*.println("Ending Topology......"); >> >> Utils.*sleep*(10000); >> >> cluster.killTopology("test"); >> >> cluster.shutdown(); >> >> } >> >> >> isolates the following info/error: >> >> >> Starting Topology...... >> >> 48124 [main] WARN o.a.s.u.Utils - STORM-VERSION new 1.2.2 old null >> >> 48129 [main] WARN o.a.s.u.Utils - STORM-VERSION new 1.2.2 old 1.2.2 >> >> 48166 [main] INFO o.a.s.d.nimbus - Received topology submission for >> test (storm-1.2.2 JDK-1.8.0_211) with conf {"topology.max.task.parallelism" >> nil, "topology.submitter.principal" "", "topology.acker.executors" nil, >> "topology.eventlogger.executors" 0, "topology.debug" true, >> "storm.zookeeper.superACL" nil, "topology.users" (), >> "topology.submitter.user" "david", "topology.kryo.register" nil, >> "topology.kryo.decorators" (), "storm.id" "test-1-1556727673", " >> topology.name" "test"} >> >> 48172 [main] INFO o.a.s.d.nimbus - uploadedJar >> >> 48183 [ProcessThread(sid:0 cport:-1):] INFO >> o.a.s.s.o.a.z.s.PrepRequestProcessor >> - Got user-level KeeperException when processing >> sessionid:0x16a7432a8950000 type:create cxid:0xb zxid:0x25 txntype:-1 >> reqpath:n/a Error Path:/storm/blobstoremaxkeysequencenumber >> Error:KeeperErrorCode = NoNode for /storm/blobstoremaxkeysequencenumber >> >> 48189 [main] INFO o.a.s.cluster - >> setup-path/blobstore/test-1-1556727673-stormconf.ser/192.168.241.5:6627-1 >> >> 48207 [main] INFO o.a.s.cluster - >> setup-path/blobstore/test-1-1556727673-stormcode.ser/192.168.241.5:6627-1 >> >> 48215 [main] INFO o.a.s.d.nimbus - desired replication count 1 >> achieved, current-replication-count for conf key = 1, >> current-replication-count for code key = 1, current-replication-count for >> jar key = 1 >> >> 48258 [main] INFO o.a.s.d.nimbus - Activating test: test-1-1556727673 >> >> Ending Topology...... >> >> >> >> I've searched, and it looks like information about the processing should >> appear here instead. I haven't found anything specific to Storm concerning >> this error. I have found references to it for Zookeeper, so tried >> installing and running Zookeeper and Kafka, with no improvement; the Storm >> tutorial didn't state this was required, but I thought I'd try anyway. At >> this point, I've exhausted my guesses, and was hoping someone could help >> point me in the proper direction. I appreciate any suggestions... >> >> >> Dave H. >> >> >> >>
