Sure, thanks.

On Mon, Aug 11, 2014 at 6:22 PM, Yan Fang <[email protected]> wrote:

> Hi Telles,
>
> I am not sure whether exporting the CLASSPATH works. (sometimes it does not
> work for me...) My suggestion is to include the hdfs jar explicitly in the
> package that you upload to hdfs. Also , remember to put the jar into your
> local samza (which is deploy/samza/lib if you go with the hello-samza
> tutorial) Let me know if that works.
>
> Cheers,
>
> Fang, Yan
> [email protected]
> +1 (206) 849-4108
>
>
> On Mon, Aug 11, 2014 at 2:04 PM, Chris Riccomini <
> [email protected]> wrote:
>
> > Hey Telles,
> >
> > Hmm. I'm out of ideas. If Zhijie is around, he'd probably be of use, but
> I
> > haven't heard from him in a while.
> >
> > I'm afraid your best bet is probably to email the YARN dev mailing list,
> > since this is a YARN config issue.
> >
> > Cheers,
> > Chris
> >
> > On 8/11/14 1:58 PM, "Telles Nobrega" <[email protected]> wrote:
> >
> > >​I exported ​export
> >
> >CLASSPATH=$CLASSPATH:hadoop-2.3.0/share/hadoop/hdfs/hadoop-hdfs-2.3.0.jar
> > >and still happened the same problem.
> > >
> > >
> > >On Mon, Aug 11, 2014 at 5:35 PM, Chris Riccomini <
> > >[email protected]> wrote:
> > >
> > >> Hey Telles,
> > >>
> > >> It sounds like either the HDFS jar is missing from the classpath, or
> the
> > >> hdfs file system needs to be configured:
> > >>
> > >> <property>
> > >>   <name>fs.hdfs.impl</name>
> > >>   <value>org.apache.hadoop.hdfs.DistributedFileSystem</value>
> > >>   <description>The FileSystem for hdfs: uris.</description>
> > >> </property>
> > >>
> > >>
> > >> (from
> > >>
> > >>
> >
> https://groups.google.com/a/cloudera.org/forum/#!topic/scm-users/lyho8ptA
> > >>zE
> > >> 0)
> > >>
> > >> I believe this will need to be configured for your NM.
> > >>
> > >> Cheers,
> > >> Chris
> > >>
> > >> On 8/11/14 1:31 PM, "Telles Nobrega" <[email protected]> wrote:
> > >>
> > >> >Yes, it is like this:
> > >> >
> > >> ><configuration>
> > >> >  <property>
> > >> >    <name>dfs.datanode.data.dir</name>
> > >> >    <value>file:///home/ubuntu/hadoop-2.3.0/hdfs/datanode</value>
> > >> >    <description>Comma separated list of paths on the local
> filesystem
> > >>of
> > >> >a
> > >> >DataNode where it should store its blocks.</description>
> > >> >  </property>
> > >> >
> > >> >  <property>
> > >> >    <name>dfs.namenode.name.dir</name>
> > >> >    <value>file:///home/ubuntu/hadoop-2.3.0/hdfs/namenode</value>
> > >> >    <description>Path on the local filesystem where the NameNode
> stores
> > >> >the
> > >> >namespace and transaction logs persistently.</description>
> > >> >  </property>
> > >> ></configuration>
> > >> >~
> > >> >
> > >> >I saw some report that this may be a classpath problem. Does this
> > >>sounds
> > >> >right to you?
> > >> >
> > >> >
> > >> >On Mon, Aug 11, 2014 at 5:25 PM, Yan Fang <[email protected]>
> > wrote:
> > >> >
> > >> >> Hi Telles,
> > >> >>
> > >> >> It looks correct. Did you put the hdfs-site.xml into your
> > >> >>HADOOP_CONF_DIR
> > >> >> ?(such as ~/.samza/conf)
> > >> >>
> > >> >> Fang, Yan
> > >> >> [email protected]
> > >> >> +1 (206) 849-4108
> > >> >>
> > >> >>
> > >> >> On Mon, Aug 11, 2014 at 1:02 PM, Telles Nobrega
> > >> >><[email protected]>
> > >> >> wrote:
> > >> >>
> > >> >> > ​Hi Yan Fang,
> > >> >> >
> > >> >> > I was able to deploy the file to hdfs, I can see them in all my
> > >>nodes
> > >> >>but
> > >> >> > when I tried running I got this error:
> > >> >> >
> > >> >> > Exception in thread "main" java.io.IOException: No FileSystem for
> > >> >>scheme:
> > >> >> > hdfs
> > >> >> > at
> > >> >>
> >
> >>org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2421)
> > >> >> >  at
> > >> >>
> > >>org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2428)
> > >> >> > at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88)
> > >> >> >  at
> > >> >>
> > >>org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467)
> > >> >> > at
> org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449)
> > >> >> >  at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367)
> > >> >> > at org.apache.hadoop.fs.Path.getFileSystem(Path.java:287)
> > >> >> >  at
> > >> >> >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>org.apache.samza.job.yarn.ClientHelper.submitApplication(ClientHelper.s
> > >>>>ca
> > >> >>la:111)
> > >> >> > at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:55)
> > >> >> >  at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:48)
> > >> >> > at org.apache.samza.job.JobRunner.run(JobRunner.scala:62)
> > >> >> >  at org.apache.samza.job.JobRunner$.main(JobRunner.scala:37)
> > >> >> > at org.apache.samza.job.JobRunner.main(JobRunner.scala)
> > >> >> >
> > >> >> >
> > >> >> > This is my yarn.package.path config:
> > >> >> >
> > >> >> >
> > >> >> >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>​yarn.package.path=hdfs://telles-master-samza:50070/samza-job-package-0
> > >>>>.7
> > >> >>.0-dist.tar.gz
> > >> >> >
> > >> >> > Thanks in advance
> > >> >> >
> > >> >> >
> > >> >> >
> > >> >> >
> > >> >> >
> > >> >> > On Mon, Aug 11, 2014 at 3:00 PM, Yan Fang <[email protected]>
> > >> >>wrote:
> > >> >> >
> > >> >> > > Hi Telles,
> > >> >> > >
> > >> >> > > In terms of "*I tried pushing the tar file to HDFS but I got an
> > >> >>error
> > >> >> > from
> > >> >> > > hadoop saying that it couldn’t find core-site.xml file*.", I
> > >>guess
> > >> >>you
> > >> >> > set
> > >> >> > > the HADOOP_CONF_DIR variable and made it point to
> ~/.samza/conf.
> > >>You
> > >> >> can
> > >> >> > do
> > >> >> > > 1) make the HADOOP_CONF_DIR point to the directory where your
> > >>conf
> > >> >> files
> > >> >> > > are, such as /etc/hadoop/conf. Or 2) copy the config files to
> > >> >> > > ~/.samza/conf. Thank you,
> > >> >> > >
> > >> >> > > Cheer,
> > >> >> > >
> > >> >> > > Fang, Yan
> > >> >> > > [email protected]
> > >> >> > > +1 (206) 849-4108
> > >> >> > >
> > >> >> > >
> > >> >> > > On Mon, Aug 11, 2014 at 7:40 AM, Chris Riccomini <
> > >> >> > > [email protected]> wrote:
> > >> >> > >
> > >> >> > > > Hey Telles,
> > >> >> > > >
> > >> >> > > > To get YARN working with the HTTP file system, you need to
> > >>follow
> > >> >>the
> > >> >> > > > instructions on:
> > >> >> > > >
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >> >>
> > >>
> > >>
> >
> http://samza.incubator.apache.org/learn/tutorials/0.7.0/run-in-multi-node
> > >> >>-y
> > >> >> > > > arn.html
> > >> >> > > >
> > >> >> > > >
> > >> >> > > > In the "Set Up Http Filesystem for YARN" section.
> > >> >> > > >
> > >> >> > > > You shouldn't need to compile anything (no Gradle, which is
> > >>what
> > >> >>your
> > >> >> > > > stack trace is showing). This setup should be done for all of
> > >>the
> > >> >> NMs,
> > >> >> > > > since they will be the ones downloading your job's package
> > >>(from
> > >> >> > > > yarn.package.path).
> > >> >> > > >
> > >> >> > > > Cheers,
> > >> >> > > > Chris
> > >> >> > > >
> > >> >> > > > On 8/9/14 9:44 PM, "Telles Nobrega" <[email protected]
> >
> > >> >>wrote:
> > >> >> > > >
> > >> >> > > > >Hi again, I tried installing the scala libs but the Http
> > >>problem
> > >> >> still
> > >> >> > > > >occurs. I realised that I need to compile incubator samza in
> > >>the
> > >> >> > > machines
> > >> >> > > > >that I¹m going to run the jobs, but the compilation fails
> with
> > >> >>this
> > >> >> > huge
> > >> >> > > > >message:
> > >> >> > > > >
> > >> >> > > > >#
> > >> >> > > > ># There is insufficient memory for the Java Runtime
> > >>Environment
> > >> >>to
> > >> >> > > > >continue.
> > >> >> > > > ># Native memory allocation (malloc) failed to allocate
> > >>3946053632
> > >> >> > bytes
> > >> >> > > > >for committing reserved memory.
> > >> >> > > > ># An error report file with more information is saved as:
> > >> >> > > > >#
> /home/ubuntu/incubator-samza/samza-kafka/hs_err_pid2506.log
> > >> >> > > > >Could not write standard input into: Gradle Worker 13.
> > >> >> > > > >java.io.IOException: Broken pipe
> > >> >> > > > >       at java.io.FileOutputStream.writeBytes(Native Method)
> > >> >> > > > >       at
> > >> >>java.io.FileOutputStream.write(FileOutputStream.java:345)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >>
> > >>java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
> > >> >> > > > >       at
> > >> >> > > >
> > >>java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.process.internal.streams.ExecOutputHandleRunner.run(ExecOut
> > >>>>>pu
> > >> >>>tH
> > >> >> > > > >andleRunner.java:53)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecuto
> > >>>>>rI
> > >> >>>mp
> > >> >> > > > >l$1.run(DefaultExecutorFactory.java:66)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j
> > >>>>>av
> > >> >>>a:
> > >> >> > > > >1145)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.
> > >>>>>ja
> > >> >>>va
> > >> >> > > > >:615)
> > >> >> > > > >       at java.lang.Thread.run(Thread.java:744)
> > >> >> > > > >Process 'Gradle Worker 13' finished with non-zero exit
> value 1
> > >> >> > > > >org.gradle.process.internal.ExecException: Process 'Gradle
> > >>Worker
> > >> >> 13'
> > >> >> > > > >finished with non-zero exit value 1
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.process.internal.DefaultExecHandle$ExecResultImpl.assertNor
> > >>>>>ma
> > >> >>>lE
> > >> >> > > > >xitValue(DefaultExecHandle.java:362)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.process.internal.DefaultWorkerProcess.onProcessStop(Default
> > >>>>>Wo
> > >> >>>rk
> > >> >> > > > >erProcess.java:89)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.process.internal.DefaultWorkerProcess.access$000(DefaultWor
> > >>>>>ke
> > >> >>>rP
> > >> >> > > > >rocess.java:33)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.process.internal.DefaultWorkerProcess$1.executionFinished(D
> > >>>>>ef
> > >> >>>au
> > >> >> > > > >ltWorkerProcess.java:55)
> > >> >> > > > >       at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> > >> >> Method)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j
> > >>>>>av
> > >> >>>a:
> > >> >> > > > >57)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccess
> > >>>>>or
> > >> >>>Im
> > >> >> > > > >pl.java:43)
> > >> >> > > > >       at java.lang.reflect.Method.invoke(Method.java:606)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDi
> > >>>>>sp
> > >> >>>at
> > >> >> > > > >ch.java:35)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDi
> > >>>>>sp
> > >> >>>at
> > >> >> > > > >ch.java:24)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java:
> > >>>>>81
> > >> >>>)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java:
> > >>>>>30
> > >> >>>)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.messaging.dispatch.ProxyDispatchAdapter$DispatchingInvocati
> > >>>>>on
> > >> >>>Ha
> > >> >> > > > >ndler.invoke(ProxyDispatchAdapter.java:93)
> > >> >> > > > >       at com.sun.proxy.$Proxy46.executionFinished(Unknown
> > >> >>Source)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.process.internal.DefaultExecHandle.setEndStateInfo(DefaultE
> > >>>>>xe
> > >> >>>cH
> > >> >> > > > >andle.java:212)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.process.internal.DefaultExecHandle.finished(DefaultExecHand
> > >>>>>le
> > >> >>>.j
> > >> >> > > > >ava:309)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.process.internal.ExecHandleRunner.completed(ExecHandleRunne
> > >>>>>r.
> > >> >>>ja
> > >> >> > > > >va:108)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.process.internal.ExecHandleRunner.run(ExecHandleRunner.java
> > >>>>>:8
> > >> >>>8)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecuto
> > >>>>>rI
> > >> >>>mp
> > >> >> > > > >l$1.run(DefaultExecutorFactory.java:66)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j
> > >>>>>av
> > >> >>>a:
> > >> >> > > > >1145)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.
> > >>>>>ja
> > >> >>>va
> > >> >> > > > >:615)
> > >> >> > > > >       at java.lang.Thread.run(Thread.java:744)
> > >> >> > > > >OpenJDK 64-Bit Server VM warning: INFO:
> > >> >> > > > >os::commit_memory(0x000000070a6c0000, 3946053632, 0) failed;
> > >> >> > > > >error='Cannot allocate memory' (errno=12)
> > >> >> > > > >#
> > >> >> > > > ># There is insufficient memory for the Java Runtime
> > >>Environment
> > >> >>to
> > >> >> > > > >continue.
> > >> >> > > > ># Native memory allocation (malloc) failed to allocate
> > >>3946053632
> > >> >> > bytes
> > >> >> > > > >for committing reserved memory.
> > >> >> > > > ># An error report file with more information is saved as:
> > >> >> > > > >#
> /home/ubuntu/incubator-samza/samza-kafka/hs_err_pid2518.log
> > >> >> > > > >Could not write standard input into: Gradle Worker 14.
> > >> >> > > > >java.io.IOException: Broken pipe
> > >> >> > > > >       at java.io.FileOutputStream.writeBytes(Native Method)
> > >> >> > > > >       at
> > >> >>java.io.FileOutputStream.write(FileOutputStream.java:345)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >>
> > >>java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
> > >> >> > > > >       at
> > >> >> > > >
> > >>java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.process.internal.streams.ExecOutputHandleRunner.run(ExecOut
> > >>>>>pu
> > >> >>>tH
> > >> >> > > > >andleRunner.java:53)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecuto
> > >>>>>rI
> > >> >>>mp
> > >> >> > > > >l$1.run(DefaultExecutorFactory.java:66)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j
> > >>>>>av
> > >> >>>a:
> > >> >> > > > >1145)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.
> > >>>>>ja
> > >> >>>va
> > >> >> > > > >:615)
> > >> >> > > > >       at java.lang.Thread.run(Thread.java:744)
> > >> >> > > > >Process 'Gradle Worker 14' finished with non-zero exit
> value 1
> > >> >> > > > >org.gradle.process.internal.ExecException: Process 'Gradle
> > >>Worker
> > >> >> 14'
> > >> >> > > > >finished with non-zero exit value 1
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.process.internal.DefaultExecHandle$ExecResultImpl.assertNor
> > >>>>>ma
> > >> >>>lE
> > >> >> > > > >xitValue(DefaultExecHandle.java:362)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.process.internal.DefaultWorkerProcess.onProcessStop(Default
> > >>>>>Wo
> > >> >>>rk
> > >> >> > > > >erProcess.java:89)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.process.internal.DefaultWorkerProcess.access$000(DefaultWor
> > >>>>>ke
> > >> >>>rP
> > >> >> > > > >rocess.java:33)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.process.internal.DefaultWorkerProcess$1.executionFinished(D
> > >>>>>ef
> > >> >>>au
> > >> >> > > > >ltWorkerProcess.java:55)
> > >> >> > > > >       at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> > >> >> Method)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j
> > >>>>>av
> > >> >>>a:
> > >> >> > > > >57)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccess
> > >>>>>or
> > >> >>>Im
> > >> >> > > > >pl.java:43)
> > >> >> > > > >       at java.lang.reflect.Method.invoke(Method.java:606)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDi
> > >>>>>sp
> > >> >>>at
> > >> >> > > > >ch.java:35)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDi
> > >>>>>sp
> > >> >>>at
> > >> >> > > > >ch.java:24)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java:
> > >>>>>81
> > >> >>>)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java:
> > >>>>>30
> > >> >>>)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.messaging.dispatch.ProxyDispatchAdapter$DispatchingInvocati
> > >>>>>on
> > >> >>>Ha
> > >> >> > > > >ndler.invoke(ProxyDispatchAdapter.java:93)
> > >> >> > > > >       at com.sun.proxy.$Proxy46.executionFinished(Unknown
> > >> >>Source)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.process.internal.DefaultExecHandle.setEndStateInfo(DefaultE
> > >>>>>xe
> > >> >>>cH
> > >> >> > > > >andle.java:212)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.process.internal.DefaultExecHandle.finished(DefaultExecHand
> > >>>>>le
> > >> >>>.j
> > >> >> > > > >ava:309)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.process.internal.ExecHandleRunner.completed(ExecHandleRunne
> > >>>>>r.
> > >> >>>ja
> > >> >> > > > >va:108)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.process.internal.ExecHandleRunner.run(ExecHandleRunner.java
> > >>>>>:8
> > >> >>>8)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecuto
> > >>>>>rI
> > >> >>>mp
> > >> >> > > > >l$1.run(DefaultExecutorFactory.java:66)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j
> > >>>>>av
> > >> >>>a:
> > >> >> > > > >1145)
> > >> >> > > > >       at
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.
> > >>>>>ja
> > >> >>>va
> > >> >> > > > >:615)
> > >> >> > > > >       at java.lang.Thread.r
> > >> >> > > > >
> > >> >> > > > >Do I need more memory for my machines? Each already has
> 4GB. I
> > >> >> really
> > >> >> > > > >need to have this running. I¹m not sure which way is best
> > >>http or
> > >> >> hdfs
> > >> >> > > > >which one you suggest and how can i solve my problem for
> each
> > >> >>case.
> > >> >> > > > >
> > >> >> > > > >Thanks in advance and sorry for bothering this much.
> > >> >> > > > >On 10 Aug 2014, at 00:20, Telles Nobrega
> > >> >><[email protected]>
> > >> >> > > wrote:
> > >> >> > > > >
> > >> >> > > > >> Hi Chris, now I have the tar file in my RM machine, and
> the
> > >> >>yarn
> > >> >> > path
> > >> >> > > > >>points to it. I changed the core-site.xml to use
> > >>HttpFileSystem
> > >> >> > instead
> > >> >> > > > >>of HDFS now it is failing with
> > >> >> > > > >>
> > >> >> > > > >> Application application_1407640485281_0001 failed 2 times
> > >>due
> > >> >>to
> > >> >> AM
> > >> >> > > > >>Container for appattempt_1407640485281_0001_000002 exited
> > >>with
> > >> >> > > > >>exitCode:-1000 due to: java.lang.ClassNotFoundException:
> > >>Class
> > >> >> > > > >>org.apache.samza.util.hadoop.HttpFileSystem not found
> > >> >> > > > >>
> > >> >> > > > >> I think I can solve this just installing scala files from
> > >>the
> > >> >> samza
> > >> >> > > > >>tutorial, can you confirm that?
> > >> >> > > > >>
> > >> >> > > > >> On 09 Aug 2014, at 08:34, Telles Nobrega
> > >> >><[email protected]
> > >> >> >
> > >> >> > > > >>wrote:
> > >> >> > > > >>
> > >> >> > > > >>> Hi Chris,
> > >> >> > > > >>>
> > >> >> > > > >>> I think the problem is that I forgot to update the
> > >> >> > yarn.job.package.
> > >> >> > > > >>> I will try again to see if it works now.
> > >> >> > > > >>>
> > >> >> > > > >>> I have one more question, how can I stop (command line)
> the
> > >> >>jobs
> > >> >> > > > >>>running in my topology, for the experiment that I will
> run,
> > >>I
> > >> >>need
> > >> >> > to
> > >> >> > > > >>>run the same job in 4 minutes intervals. So I need to kill
> > >>it,
> > >> >> clean
> > >> >> > > > >>>the kafka topics and rerun.
> > >> >> > > > >>>
> > >> >> > > > >>> Thanks in advance.
> > >> >> > > > >>>
> > >> >> > > > >>> On 08 Aug 2014, at 12:41, Chris Riccomini
> > >> >> > > > >>><[email protected]> wrote:
> > >> >> > > > >>>
> > >> >> > > > >>>> Hey Telles,
> > >> >> > > > >>>>
> > >> >> > > > >>>>>> Do I need to have the job folder on each machine in my
> > >> >> cluster?
> > >> >> > > > >>>>
> > >> >> > > > >>>> No, you should not need to do this. There are two ways
> to
> > >> >>deploy
> > >> >> > > your
> > >> >> > > > >>>> tarball to the YARN grid. One is to put it in HDFS, and
> > >>the
> > >> >> other
> > >> >> > is
> > >> >> > > > >>>>to
> > >> >> > > > >>>> put it on an HTTP server. The link to running a Samza
> job
> > >>in
> > >> >>a
> > >> >> > > > >>>>multi-node
> > >> >> > > > >>>> YARN cluster describes how to do both (either HTTP
> server
> > >>or
> > >> >> > HDFS).
> > >> >> > > > >>>>
> > >> >> > > > >>>> In both cases, once the tarball is put in on the
> HTTP/HDFS
> > >> >> > > server(s),
> > >> >> > > > >>>>you
> > >> >> > > > >>>> must update yarn.package.path to point to it. From
> there,
> > >>the
> > >> >> YARN
> > >> >> > > NM
> > >> >> > > > >>>> should download it for you automatically when you start
> > >>your
> > >> >> job.
> > >> >> > > > >>>>
> > >> >> > > > >>>> * Can you send along a paste of your job config?
> > >> >> > > > >>>>
> > >> >> > > > >>>> Cheers,
> > >> >> > > > >>>> Chris
> > >> >> > > > >>>>
> > >> >> > > > >>>> On 8/8/14 8:04 AM, "Claudio Martins"
> > >> >><[email protected]>
> > >> >> > > wrote:
> > >> >> > > > >>>>
> > >> >> > > > >>>>> Hi Telles, it looks to me that you forgot to update the
> > >> >> > > > >>>>> "yarn.package.path"
> > >> >> > > > >>>>> attribute in your config file for the task.
> > >> >> > > > >>>>>
> > >> >> > > > >>>>> - Claudio Martins
> > >> >> > > > >>>>> Head of Engineering
> > >> >> > > > >>>>> MobileAware USA Inc. / www.mobileaware.com
> > >> >> > > > >>>>> office: +1 617 986 5060 / mobile: +1 617 480 5288
> > >> >> > > > >>>>> linkedin: www.linkedin.com/in/martinsclaudio
> > >> >> > > > >>>>>
> > >> >> > > > >>>>>
> > >> >> > > > >>>>> On Fri, Aug 8, 2014 at 10:55 AM, Telles Nobrega
> > >> >> > > > >>>>><[email protected]>
> > >> >> > > > >>>>> wrote:
> > >> >> > > > >>>>>
> > >> >> > > > >>>>>> Hi,
> > >> >> > > > >>>>>>
> > >> >> > > > >>>>>> this is my first time trying to run a job on a
> multinode
> > >> >> > > > >>>>>>environment. I
> > >> >> > > > >>>>>> have the cluster set up, I can see in the GUI that all
> > >> >>nodes
> > >> >> are
> > >> >> > > > >>>>>> working.
> > >> >> > > > >>>>>> Do I need to have the job folder on each machine in my
> > >> >> cluster?
> > >> >> > > > >>>>>> - The first time I tried running with the job on the
> > >> >>namenode
> > >> >> > > > >>>>>>machine
> > >> >> > > > >>>>>> and
> > >> >> > > > >>>>>> it failed saying:
> > >> >> > > > >>>>>>
> > >> >> > > > >>>>>> Application application_1407509228798_0001 failed 2
> > >>times
> > >> >>due
> > >> >> to
> > >> >> > > AM
> > >> >> > > > >>>>>> Container for appattempt_1407509228798_0001_000002
> > >>exited
> > >> >>with
> > >> >> > > > >>>>>>exitCode:
> > >> >> > > > >>>>>> -1000 due to: File
> > >> >> > > > >>>>>>
> > >> >> > > > >>>>>>
> > >> >> > > > >>>>>>
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>>>>>>file:/home/ubuntu/alarm-samza/samza-job-package/target/samza-job-
> > >>>>>>>>>>pa
> > >> >>>>>>>>ck
> > >> >> > > > >>>>>>age-
> > >> >> > > > >>>>>> 0.7.0-dist.tar.gz
> > >> >> > > > >>>>>> does not exist
> > >> >> > > > >>>>>>
> > >> >> > > > >>>>>> So I copied the folder to each machine in my cluster
> and
> > >> >>got
> > >> >> > this
> > >> >> > > > >>>>>>error:
> > >> >> > > > >>>>>>
> > >> >> > > > >>>>>> Application application_1407509228798_0002 failed 2
> > >>times
> > >> >>due
> > >> >> to
> > >> >> > > AM
> > >> >> > > > >>>>>> Container for appattempt_1407509228798_0002_000002
> > >>exited
> > >> >>with
> > >> >> > > > >>>>>>exitCode:
> > >> >> > > > >>>>>> -1000 due to: Resource
> > >> >> > > > >>>>>>
> > >> >> > > > >>>>>>
> > >> >> > > > >>>>>>
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >>
> > >>
> >
> >>>>>>>>>>file:/home/ubuntu/alarm-samza/samza-job-package/target/samza-job-
> > >>>>>>>>>>pa
> > >> >>>>>>>>ck
> > >> >> > > > >>>>>>age-
> > >> >> > > > >>>>>> 0.7.0-dist.tar.gz
> > >> >> > > > >>>>>> changed on src filesystem (expected 1407509168000, was
> > >> >> > > 1407509434000
> > >> >> > > > >>>>>>
> > >> >> > > > >>>>>> What am I missing?
> > >> >> > > > >>>>>>
> > >> >> > > > >>>>>> p.s.: I followed this
> > >> >> > > > >>>>>>
> > >> >> > > > >>>>>><
> > >> >> > > >
> > >> >>
> https://github.com/yahoo/samoa/wiki/Executing-SAMOA-with-Apache-Samz
> > >> >> > > > >>>>>>a>
> > >> >> > > > >>>>>> tutorial
> > >> >> > > > >>>>>> and this
> > >> >> > > > >>>>>> <
> > >> >> > > > >>>>>>
> > >> >> > > > >>>>>>
> > >> >> > > > >>>>>>
> > >> >> > > >
> > >> >>
> > http://samza.incubator.apache.org/learn/tutorials/0.7.0/run-in-multi-
> > >> >> > > > >>>>>>node
> > >> >> > > > >>>>>> -yarn.html
> > >> >> > > > >>>>>>>
> > >> >> > > > >>>>>> to
> > >> >> > > > >>>>>> set up the cluster.
> > >> >> > > > >>>>>>
> > >> >> > > > >>>>>> Help is much appreciated.
> > >> >> > > > >>>>>>
> > >> >> > > > >>>>>> Thanks in advance.
> > >> >> > > > >>>>>>
> > >> >> > > > >>>>>> --
> > >> >> > > > >>>>>> ------------------------------------------
> > >> >> > > > >>>>>> Telles Mota Vidal Nobrega
> > >> >> > > > >>>>>> M.sc. Candidate at UFCG
> > >> >> > > > >>>>>> B.sc. in Computer Science at UFCG
> > >> >> > > > >>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG
> > >> >> > > > >>>>>>
> > >> >> > > > >>>>
> > >> >> > > > >>>
> > >> >> > > > >>
> > >> >> > > > >
> > >> >> > > >
> > >> >> > > >
> > >> >> > >
> > >> >> >
> > >> >> >
> > >> >> >
> > >> >> > --
> > >> >> > ------------------------------------------
> > >> >> > Telles Mota Vidal Nobrega
> > >> >> > M.sc. Candidate at UFCG
> > >> >> > B.sc. in Computer Science at UFCG
> > >> >> > Software Engineer at OpenStack Project - HP/LSD-UFCG
> > >> >> >
> > >> >>
> > >> >
> > >> >
> > >> >
> > >> >--
> > >> >------------------------------------------
> > >> >Telles Mota Vidal Nobrega
> > >> >M.sc. Candidate at UFCG
> > >> >B.sc. in Computer Science at UFCG
> > >> >Software Engineer at OpenStack Project - HP/LSD-UFCG
> > >>
> > >>
> > >
> > >
> > >--
> > >------------------------------------------
> > >Telles Mota Vidal Nobrega
> > >M.sc. Candidate at UFCG
> > >B.sc. in Computer Science at UFCG
> > >Software Engineer at OpenStack Project - HP/LSD-UFCG
> >
> >
>



-- 
------------------------------------------
Telles Mota Vidal Nobrega
M.sc. Candidate at UFCG
B.sc. in Computer Science at UFCG
Software Engineer at OpenStack Project - HP/LSD-UFCG

Reply via email to