Opps, it seems that the hadoop-hdfs-2.3.0.jar lost some dependencies. It's
a hadoop-specific exception.

* Could you add the jar file that contains the missing class? (not quite
sure it's in hadoop-common or hadoop-client)

* Another worth trying is to add the dependency from pom file, let maven
take care of it. Assume you are using hello-samza similar structure. There
are two steps:
  ** add dependency in pom.xml of samza-job-package
  ** add <include>org.apache.hadoop:hadoop-hdfs</include> in
samza-job-package/src/main/assembly/src.xml
Then mvn clean package.

Thank you.

Cheers,

Fang, Yan
[email protected]
+1 (206) 849-4108


On Mon, Aug 11, 2014 at 5:12 PM, Telles Nobrega <[email protected]>
wrote:

> Same thing happens.
>
> On 11 Aug 2014, at 21:05, Yan Fang <[email protected]> wrote:
>
> > Cool, we are almost there. Could you remove
> >
> > <property>
> >  <name>fs.hdfs.impl</name>
> >  <value>org.apache.hadoop.hdfs.
> > DistributedFileSystem</value>
> >  <description>The FileSystem for hdfs: uris.</description>
> > </property>
> >
> > To see how it works?
> >
> >
> > Fang, Yan
> > [email protected]
> > +1 (206) 849-4108
> >
> >
> > On Mon, Aug 11, 2014 at 5:03 PM, Telles Nobrega <[email protected]
> >
> > wrote:
> >
> >> You may forget this last email, I was really stupid and put the files
> in a
> >> different folder. Now it could find the file but it’s not there yet…
> >> another error came up
> >>
> >> Exception in thread "main" java.util.ServiceConfigurationError:
> >> org.apache.hadoop.fs.FileSystem: Provider
> >> org.apache.hadoop.hdfs.DistributedFileSystem could not be instantiated
> >>        at java.util.ServiceLoader.fail(ServiceLoader.java:224)
> >>        at java.util.ServiceLoader.access$100(ServiceLoader.java:181)
> >>        at
> >> java.util.ServiceLoader$LazyIterator.next(ServiceLoader.java:377)
> >>        at java.util.ServiceLoader$1.next(ServiceLoader.java:445)
> >>        at
> >> org.apache.hadoop.fs.FileSystem.loadFileSystems(FileSystem.java:2400)
> >>        at
> >> org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2411)
> >>        at
> >> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2428)
> >>        at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88)
> >>        at
> >> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467)
> >>        at
> org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449)
> >>        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367)
> >>        at org.apache.hadoop.fs.Path.getFileSystem(Path.java:287)
> >>        at
> >>
> org.apache.samza.job.yarn.ClientHelper.submitApplication(ClientHelper.scala:111)
> >>        at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:55)
> >>        at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:48)
> >>        at org.apache.samza.job.JobRunner.run(JobRunner.scala:62)
> >>        at org.apache.samza.job.JobRunner$.main(JobRunner.scala:37)
> >>        at org.apache.samza.job.JobRunner.main(JobRunner.scala)
> >> Caused by: java.lang.NoClassDefFoundError:
> >> org/apache/hadoop/conf/Configuration$DeprecationDelta
> >>        at
> >>
> org.apache.hadoop.hdfs.HdfsConfiguration.addDeprecatedKeys(HdfsConfiguration.java:66)
> >>        at
> >>
> org.apache.hadoop.hdfs.HdfsConfiguration.<clinit>(HdfsConfiguration.java:31)
> >>        at
> >>
> org.apache.hadoop.hdfs.DistributedFileSystem.<clinit>(DistributedFileSystem.java:106)
> >>        at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
> >> Method)
> >>        at
> >>
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
> >>        at
> >>
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> >>        at
> java.lang.reflect.Constructor.newInstance(Constructor.java:526)
> >>        at java.lang.Class.newInstance(Class.java:374)
> >>        at
> >> java.util.ServiceLoader$LazyIterator.next(ServiceLoader.java:373)
> >>        ... 15 more
> >> Caused by: java.lang.ClassNotFoundException:
> >> org.apache.hadoop.conf.Configuration$DeprecationDelta
> >>        at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
> >>        at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
> >>        at java.security.AccessController.doPrivileged(Native Method)
> >>        at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
> >>        at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
> >>        at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
> >>        at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
> >>        ... 24 more
> >>
> >> On 11 Aug 2014, at 20:45, Telles Nobrega <[email protected]>
> wrote:
> >>
> >>> Hi, I copied hadoop-hdfs-2.3.0 to my-job/lib and it changed the error
> >> which is good but the error is back to
> >>>
> >>> Exception in thread "main" java.lang.RuntimeException:
> >> java.lang.ClassNotFoundException: Class
> >> org.apache.hadoop.hdfs.DistributedFileSystem not found
> >>>      at
> >> org.apache.hadoop.conf.Configuration.getClass(Configuration.java:1720)
> >>>      at
> >> org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2415)
> >>>      at
> >> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2428)
> >>>      at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88)
> >>>      at
> >> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467)
> >>>      at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449)
> >>>      at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367)
> >>>      at org.apache.hadoop.fs.Path.getFileSystem(Path.java:287)
> >>>      at
> >>
> org.apache.samza.job.yarn.ClientHelper.submitApplication(ClientHelper.scala:111)
> >>>      at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:55)
> >>>      at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:48)
> >>>      at org.apache.samza.job.JobRunner.run(JobRunner.scala:62)
> >>>      at org.apache.samza.job.JobRunner$.main(JobRunner.scala:37)
> >>>      at org.apache.samza.job.JobRunner.main(JobRunner.scala)
> >>> Caused by: java.lang.ClassNotFoundException: Class
> >> org.apache.hadoop.hdfs.DistributedFileSystem not found
> >>>      at
> >>
> org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:1626)
> >>>      at
> >> org.apache.hadoop.conf.Configuration.getClass(Configuration.java:1718)
> >>>      ... 13 more
> >>>
> >>> Do I need to have this lib in all nodes at the job folder or just to
> >> submit?
> >>>
> >>> On 11 Aug 2014, at 20:11, Yan Fang <[email protected]> wrote:
> >>>
> >>>> Hi Telles,
> >>>>
> >>>> I replayed your problem and think I figured out why CLASSPATH does not
> >>>> work. Because in our script bin/run-class.sh, we have the line
> >>>> "CLASSPATH=$HADOOP_CONF_DIR", which actually ingores your setting.
> >>>>
> >>>> So a simple solution is to copy the hadoop-hdfs.jar to your samza lib
> >>>> directory. Then run bin/run-job ----config-factory=...
> >> --config-path=... .
> >>>> Let me know how it goes. Thank you.
> >>>>
> >>>> Cheers,
> >>>>
> >>>> Fang, Yan
> >>>> [email protected]
> >>>> +1 (206) 849-4108
> >>>>
> >>>>
> >>>> On Mon, Aug 11, 2014 at 4:07 PM, Telles Nobrega <
> >> [email protected]>
> >>>> wrote:
> >>>>
> >>>>> Sure, thanks.
> >>>>>
> >>>>>
> >>>>> On Mon, Aug 11, 2014 at 6:22 PM, Yan Fang <[email protected]>
> >> wrote:
> >>>>>
> >>>>>> Hi Telles,
> >>>>>>
> >>>>>> I am not sure whether exporting the CLASSPATH works. (sometimes it
> >> does
> >>>>> not
> >>>>>> work for me...) My suggestion is to include the hdfs jar explicitly
> in
> >>>>> the
> >>>>>> package that you upload to hdfs. Also , remember to put the jar into
> >> your
> >>>>>> local samza (which is deploy/samza/lib if you go with the
> hello-samza
> >>>>>> tutorial) Let me know if that works.
> >>>>>>
> >>>>>> Cheers,
> >>>>>>
> >>>>>> Fang, Yan
> >>>>>> [email protected]
> >>>>>> +1 (206) 849-4108
> >>>>>>
> >>>>>>
> >>>>>> On Mon, Aug 11, 2014 at 2:04 PM, Chris Riccomini <
> >>>>>> [email protected]> wrote:
> >>>>>>
> >>>>>>> Hey Telles,
> >>>>>>>
> >>>>>>> Hmm. I'm out of ideas. If Zhijie is around, he'd probably be of
> use,
> >>>>> but
> >>>>>> I
> >>>>>>> haven't heard from him in a while.
> >>>>>>>
> >>>>>>> I'm afraid your best bet is probably to email the YARN dev mailing
> >>>>> list,
> >>>>>>> since this is a YARN config issue.
> >>>>>>>
> >>>>>>> Cheers,
> >>>>>>> Chris
> >>>>>>>
> >>>>>>> On 8/11/14 1:58 PM, "Telles Nobrega" <[email protected]>
> >> wrote:
> >>>>>>>
> >>>>>>>> ​I exported ​export
> >>>>>>>
> >>>>>>
> >>>>>>
> >>
> CLASSPATH=$CLASSPATH:hadoop-2.3.0/share/hadoop/hdfs/hadoop-hdfs-2.3.0.jar
> >>>>>>>> and still happened the same problem.
> >>>>>>>>
> >>>>>>>>
> >>>>>>>> On Mon, Aug 11, 2014 at 5:35 PM, Chris Riccomini <
> >>>>>>>> [email protected]> wrote:
> >>>>>>>>
> >>>>>>>>> Hey Telles,
> >>>>>>>>>
> >>>>>>>>> It sounds like either the HDFS jar is missing from the classpath,
> >> or
> >>>>>> the
> >>>>>>>>> hdfs file system needs to be configured:
> >>>>>>>>>
> >>>>>>>>> <property>
> >>>>>>>>> <name>fs.hdfs.impl</name>
> >>>>>>>>> <value>org.apache.hadoop.hdfs.DistributedFileSystem</value>
> >>>>>>>>> <description>The FileSystem for hdfs: uris.</description>
> >>>>>>>>> </property>
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>> (from
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>
> >>
> https://groups.google.com/a/cloudera.org/forum/#!topic/scm-users/lyho8ptA
> >>>>>>>>> zE
> >>>>>>>>> 0)
> >>>>>>>>>
> >>>>>>>>> I believe this will need to be configured for your NM.
> >>>>>>>>>
> >>>>>>>>> Cheers,
> >>>>>>>>> Chris
> >>>>>>>>>
> >>>>>>>>> On 8/11/14 1:31 PM, "Telles Nobrega" <[email protected]>
> >>>>> wrote:
> >>>>>>>>>
> >>>>>>>>>> Yes, it is like this:
> >>>>>>>>>>
> >>>>>>>>>> <configuration>
> >>>>>>>>>> <property>
> >>>>>>>>>> <name>dfs.datanode.data.dir</name>
> >>>>>>>>>> <value>file:///home/ubuntu/hadoop-2.3.0/hdfs/datanode</value>
> >>>>>>>>>> <description>Comma separated list of paths on the local
> >>>>>> filesystem
> >>>>>>>>> of
> >>>>>>>>>> a
> >>>>>>>>>> DataNode where it should store its blocks.</description>
> >>>>>>>>>> </property>
> >>>>>>>>>>
> >>>>>>>>>> <property>
> >>>>>>>>>> <name>dfs.namenode.name.dir</name>
> >>>>>>>>>> <value>file:///home/ubuntu/hadoop-2.3.0/hdfs/namenode</value>
> >>>>>>>>>> <description>Path on the local filesystem where the NameNode
> >>>>>> stores
> >>>>>>>>>> the
> >>>>>>>>>> namespace and transaction logs persistently.</description>
> >>>>>>>>>> </property>
> >>>>>>>>>> </configuration>
> >>>>>>>>>> ~
> >>>>>>>>>>
> >>>>>>>>>> I saw some report that this may be a classpath problem. Does
> this
> >>>>>>>>> sounds
> >>>>>>>>>> right to you?
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>> On Mon, Aug 11, 2014 at 5:25 PM, Yan Fang <[email protected]
> >
> >>>>>>> wrote:
> >>>>>>>>>>
> >>>>>>>>>>> Hi Telles,
> >>>>>>>>>>>
> >>>>>>>>>>> It looks correct. Did you put the hdfs-site.xml into your
> >>>>>>>>>>> HADOOP_CONF_DIR
> >>>>>>>>>>> ?(such as ~/.samza/conf)
> >>>>>>>>>>>
> >>>>>>>>>>> Fang, Yan
> >>>>>>>>>>> [email protected]
> >>>>>>>>>>> +1 (206) 849-4108
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>> On Mon, Aug 11, 2014 at 1:02 PM, Telles Nobrega
> >>>>>>>>>>> <[email protected]>
> >>>>>>>>>>> wrote:
> >>>>>>>>>>>
> >>>>>>>>>>>> ​Hi Yan Fang,
> >>>>>>>>>>>>
> >>>>>>>>>>>> I was able to deploy the file to hdfs, I can see them in all
> my
> >>>>>>>>> nodes
> >>>>>>>>>>> but
> >>>>>>>>>>>> when I tried running I got this error:
> >>>>>>>>>>>>
> >>>>>>>>>>>> Exception in thread "main" java.io.IOException: No FileSystem
> >>>>> for
> >>>>>>>>>>> scheme:
> >>>>>>>>>>>> hdfs
> >>>>>>>>>>>> at
> >>>>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>
> >> org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2421)
> >>>>>>>>>>>> at
> >>>>>>>>>>>
> >>>>>>>
> >>>>>>>
> >> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2428)
> >>>>>>>>>>>> at
> >>>>> org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88)
> >>>>>>>>>>>> at
> >>>>>>>>>>>
> >>>>>>>
> >>>>>>>
> >> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467)
> >>>>>>>>>>>> at
> >>>>>> org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449)
> >>>>>>>>>>>> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367)
> >>>>>>>>>>>> at org.apache.hadoop.fs.Path.getFileSystem(Path.java:287)
> >>>>>>>>>>>> at
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>
> >> org.apache.samza.job.yarn.ClientHelper.submitApplication(ClientHelper.s
> >>>>>>>>>>> ca
> >>>>>>>>>>> la:111)
> >>>>>>>>>>>> at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:55)
> >>>>>>>>>>>> at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:48)
> >>>>>>>>>>>> at org.apache.samza.job.JobRunner.run(JobRunner.scala:62)
> >>>>>>>>>>>> at org.apache.samza.job.JobRunner$.main(JobRunner.scala:37)
> >>>>>>>>>>>> at org.apache.samza.job.JobRunner.main(JobRunner.scala)
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>> This is my yarn.package.path config:
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>
> >> ​yarn.package.path=hdfs://telles-master-samza:50070/samza-job-package-0
> >>>>>>>>>>> .7
> >>>>>>>>>>> .0-dist.tar.gz
> >>>>>>>>>>>>
> >>>>>>>>>>>> Thanks in advance
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>> On Mon, Aug 11, 2014 at 3:00 PM, Yan Fang <
> >>>>> [email protected]>
> >>>>>>>>>>> wrote:
> >>>>>>>>>>>>
> >>>>>>>>>>>>> Hi Telles,
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> In terms of "*I tried pushing the tar file to HDFS but I got
> >>>>> an
> >>>>>>>>>>> error
> >>>>>>>>>>>> from
> >>>>>>>>>>>>> hadoop saying that it couldn’t find core-site.xml file*.", I
> >>>>>>>>> guess
> >>>>>>>>>>> you
> >>>>>>>>>>>> set
> >>>>>>>>>>>>> the HADOOP_CONF_DIR variable and made it point to
> >>>>>> ~/.samza/conf.
> >>>>>>>>> You
> >>>>>>>>>>> can
> >>>>>>>>>>>> do
> >>>>>>>>>>>>> 1) make the HADOOP_CONF_DIR point to the directory where your
> >>>>>>>>> conf
> >>>>>>>>>>> files
> >>>>>>>>>>>>> are, such as /etc/hadoop/conf. Or 2) copy the config files to
> >>>>>>>>>>>>> ~/.samza/conf. Thank you,
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> Cheer,
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> Fang, Yan
> >>>>>>>>>>>>> [email protected]
> >>>>>>>>>>>>> +1 (206) 849-4108
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> On Mon, Aug 11, 2014 at 7:40 AM, Chris Riccomini <
> >>>>>>>>>>>>> [email protected]> wrote:
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>> Hey Telles,
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>> To get YARN working with the HTTP file system, you need to
> >>>>>>>>> follow
> >>>>>>>>>>> the
> >>>>>>>>>>>>>> instructions on:
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>
> >>
> http://samza.incubator.apache.org/learn/tutorials/0.7.0/run-in-multi-node
> >>>>>>>>>>> -y
> >>>>>>>>>>>>>> arn.html
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>> In the "Set Up Http Filesystem for YARN" section.
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>> You shouldn't need to compile anything (no Gradle, which is
> >>>>>>>>> what
> >>>>>>>>>>> your
> >>>>>>>>>>>>>> stack trace is showing). This setup should be done for all
> >>>>> of
> >>>>>>>>> the
> >>>>>>>>>>> NMs,
> >>>>>>>>>>>>>> since they will be the ones downloading your job's package
> >>>>>>>>> (from
> >>>>>>>>>>>>>> yarn.package.path).
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>> Cheers,
> >>>>>>>>>>>>>> Chris
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>> On 8/9/14 9:44 PM, "Telles Nobrega" <
> >>>>> [email protected]
> >>>>>>>
> >>>>>>>>>>> wrote:
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>>> Hi again, I tried installing the scala libs but the Http
> >>>>>>>>> problem
> >>>>>>>>>>> still
> >>>>>>>>>>>>>>> occurs. I realised that I need to compile incubator samza
> >>>>> in
> >>>>>>>>> the
> >>>>>>>>>>>>> machines
> >>>>>>>>>>>>>>> that I¹m going to run the jobs, but the compilation fails
> >>>>>> with
> >>>>>>>>>>> this
> >>>>>>>>>>>> huge
> >>>>>>>>>>>>>>> message:
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>> #
> >>>>>>>>>>>>>>> # There is insufficient memory for the Java Runtime
> >>>>>>>>> Environment
> >>>>>>>>>>> to
> >>>>>>>>>>>>>>> continue.
> >>>>>>>>>>>>>>> # Native memory allocation (malloc) failed to allocate
> >>>>>>>>> 3946053632
> >>>>>>>>>>>> bytes
> >>>>>>>>>>>>>>> for committing reserved memory.
> >>>>>>>>>>>>>>> # An error report file with more information is saved as:
> >>>>>>>>>>>>>>> #
> >>>>>> /home/ubuntu/incubator-samza/samza-kafka/hs_err_pid2506.log
> >>>>>>>>>>>>>>> Could not write standard input into: Gradle Worker 13.
> >>>>>>>>>>>>>>> java.io.IOException: Broken pipe
> >>>>>>>>>>>>>>>    at java.io.FileOutputStream.writeBytes(Native
> >>>>> Method)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>> java.io.FileOutputStream.write(FileOutputStream.java:345)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>
> >>>>>>>
> >> java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>> java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.process.internal.streams.ExecOutputHandleRunner.run(ExecOut
> >>>>>>>>>>>> pu
> >>>>>>>>>>>> tH
> >>>>>>>>>>>>>>> andleRunner.java:53)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecuto
> >>>>>>>>>>>> rI
> >>>>>>>>>>>> mp
> >>>>>>>>>>>>>>> l$1.run(DefaultExecutorFactory.java:66)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j
> >>>>>>>>>>>> av
> >>>>>>>>>>>> a:
> >>>>>>>>>>>>>>> 1145)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.
> >>>>>>>>>>>> ja
> >>>>>>>>>>>> va
> >>>>>>>>>>>>>>> :615)
> >>>>>>>>>>>>>>>    at java.lang.Thread.run(Thread.java:744)
> >>>>>>>>>>>>>>> Process 'Gradle Worker 13' finished with non-zero exit
> >>>>>> value 1
> >>>>>>>>>>>>>>> org.gradle.process.internal.ExecException: Process 'Gradle
> >>>>>>>>> Worker
> >>>>>>>>>>> 13'
> >>>>>>>>>>>>>>> finished with non-zero exit value 1
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.process.internal.DefaultExecHandle$ExecResultImpl.assertNor
> >>>>>>>>>>>> ma
> >>>>>>>>>>>> lE
> >>>>>>>>>>>>>>> xitValue(DefaultExecHandle.java:362)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.process.internal.DefaultWorkerProcess.onProcessStop(Default
> >>>>>>>>>>>> Wo
> >>>>>>>>>>>> rk
> >>>>>>>>>>>>>>> erProcess.java:89)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.process.internal.DefaultWorkerProcess.access$000(DefaultWor
> >>>>>>>>>>>> ke
> >>>>>>>>>>>> rP
> >>>>>>>>>>>>>>> rocess.java:33)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.process.internal.DefaultWorkerProcess$1.executionFinished(D
> >>>>>>>>>>>> ef
> >>>>>>>>>>>> au
> >>>>>>>>>>>>>>> ltWorkerProcess.java:55)
> >>>>>>>>>>>>>>>    at
> >>>>>> sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> >>>>>>>>>>> Method)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j
> >>>>>>>>>>>> av
> >>>>>>>>>>>> a:
> >>>>>>>>>>>>>>> 57)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccess
> >>>>>>>>>>>> or
> >>>>>>>>>>>> Im
> >>>>>>>>>>>>>>> pl.java:43)
> >>>>>>>>>>>>>>>    at java.lang.reflect.Method.invoke(Method.java:606)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDi
> >>>>>>>>>>>> sp
> >>>>>>>>>>>> at
> >>>>>>>>>>>>>>> ch.java:35)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDi
> >>>>>>>>>>>> sp
> >>>>>>>>>>>> at
> >>>>>>>>>>>>>>> ch.java:24)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java:
> >>>>>>>>>>>> 81
> >>>>>>>>>>>> )
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java:
> >>>>>>>>>>>> 30
> >>>>>>>>>>>> )
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.messaging.dispatch.ProxyDispatchAdapter$DispatchingInvocati
> >>>>>>>>>>>> on
> >>>>>>>>>>>> Ha
> >>>>>>>>>>>>>>> ndler.invoke(ProxyDispatchAdapter.java:93)
> >>>>>>>>>>>>>>>    at com.sun.proxy.$Proxy46.executionFinished(Unknown
> >>>>>>>>>>> Source)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.process.internal.DefaultExecHandle.setEndStateInfo(DefaultE
> >>>>>>>>>>>> xe
> >>>>>>>>>>>> cH
> >>>>>>>>>>>>>>> andle.java:212)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.process.internal.DefaultExecHandle.finished(DefaultExecHand
> >>>>>>>>>>>> le
> >>>>>>>>>>>> .j
> >>>>>>>>>>>>>>> ava:309)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.process.internal.ExecHandleRunner.completed(ExecHandleRunne
> >>>>>>>>>>>> r.
> >>>>>>>>>>>> ja
> >>>>>>>>>>>>>>> va:108)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.process.internal.ExecHandleRunner.run(ExecHandleRunner.java
> >>>>>>>>>>>> :8
> >>>>>>>>>>>> 8)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecuto
> >>>>>>>>>>>> rI
> >>>>>>>>>>>> mp
> >>>>>>>>>>>>>>> l$1.run(DefaultExecutorFactory.java:66)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j
> >>>>>>>>>>>> av
> >>>>>>>>>>>> a:
> >>>>>>>>>>>>>>> 1145)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.
> >>>>>>>>>>>> ja
> >>>>>>>>>>>> va
> >>>>>>>>>>>>>>> :615)
> >>>>>>>>>>>>>>>    at java.lang.Thread.run(Thread.java:744)
> >>>>>>>>>>>>>>> OpenJDK 64-Bit Server VM warning: INFO:
> >>>>>>>>>>>>>>> os::commit_memory(0x000000070a6c0000, 3946053632, 0)
> >>>>> failed;
> >>>>>>>>>>>>>>> error='Cannot allocate memory' (errno=12)
> >>>>>>>>>>>>>>> #
> >>>>>>>>>>>>>>> # There is insufficient memory for the Java Runtime
> >>>>>>>>> Environment
> >>>>>>>>>>> to
> >>>>>>>>>>>>>>> continue.
> >>>>>>>>>>>>>>> # Native memory allocation (malloc) failed to allocate
> >>>>>>>>> 3946053632
> >>>>>>>>>>>> bytes
> >>>>>>>>>>>>>>> for committing reserved memory.
> >>>>>>>>>>>>>>> # An error report file with more information is saved as:
> >>>>>>>>>>>>>>> #
> >>>>>> /home/ubuntu/incubator-samza/samza-kafka/hs_err_pid2518.log
> >>>>>>>>>>>>>>> Could not write standard input into: Gradle Worker 14.
> >>>>>>>>>>>>>>> java.io.IOException: Broken pipe
> >>>>>>>>>>>>>>>    at java.io.FileOutputStream.writeBytes(Native
> >>>>> Method)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>> java.io.FileOutputStream.write(FileOutputStream.java:345)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>
> >>>>>>>
> >> java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>> java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.process.internal.streams.ExecOutputHandleRunner.run(ExecOut
> >>>>>>>>>>>> pu
> >>>>>>>>>>>> tH
> >>>>>>>>>>>>>>> andleRunner.java:53)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecuto
> >>>>>>>>>>>> rI
> >>>>>>>>>>>> mp
> >>>>>>>>>>>>>>> l$1.run(DefaultExecutorFactory.java:66)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j
> >>>>>>>>>>>> av
> >>>>>>>>>>>> a:
> >>>>>>>>>>>>>>> 1145)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.
> >>>>>>>>>>>> ja
> >>>>>>>>>>>> va
> >>>>>>>>>>>>>>> :615)
> >>>>>>>>>>>>>>>    at java.lang.Thread.run(Thread.java:744)
> >>>>>>>>>>>>>>> Process 'Gradle Worker 14' finished with non-zero exit
> >>>>>> value 1
> >>>>>>>>>>>>>>> org.gradle.process.internal.ExecException: Process 'Gradle
> >>>>>>>>> Worker
> >>>>>>>>>>> 14'
> >>>>>>>>>>>>>>> finished with non-zero exit value 1
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.process.internal.DefaultExecHandle$ExecResultImpl.assertNor
> >>>>>>>>>>>> ma
> >>>>>>>>>>>> lE
> >>>>>>>>>>>>>>> xitValue(DefaultExecHandle.java:362)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.process.internal.DefaultWorkerProcess.onProcessStop(Default
> >>>>>>>>>>>> Wo
> >>>>>>>>>>>> rk
> >>>>>>>>>>>>>>> erProcess.java:89)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.process.internal.DefaultWorkerProcess.access$000(DefaultWor
> >>>>>>>>>>>> ke
> >>>>>>>>>>>> rP
> >>>>>>>>>>>>>>> rocess.java:33)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.process.internal.DefaultWorkerProcess$1.executionFinished(D
> >>>>>>>>>>>> ef
> >>>>>>>>>>>> au
> >>>>>>>>>>>>>>> ltWorkerProcess.java:55)
> >>>>>>>>>>>>>>>    at
> >>>>>> sun.reflect.NativeMethodAccessorImpl.invoke0(Native
> >>>>>>>>>>> Method)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j
> >>>>>>>>>>>> av
> >>>>>>>>>>>> a:
> >>>>>>>>>>>>>>> 57)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccess
> >>>>>>>>>>>> or
> >>>>>>>>>>>> Im
> >>>>>>>>>>>>>>> pl.java:43)
> >>>>>>>>>>>>>>>    at java.lang.reflect.Method.invoke(Method.java:606)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDi
> >>>>>>>>>>>> sp
> >>>>>>>>>>>> at
> >>>>>>>>>>>>>>> ch.java:35)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDi
> >>>>>>>>>>>> sp
> >>>>>>>>>>>> at
> >>>>>>>>>>>>>>> ch.java:24)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java:
> >>>>>>>>>>>> 81
> >>>>>>>>>>>> )
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java:
> >>>>>>>>>>>> 30
> >>>>>>>>>>>> )
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.messaging.dispatch.ProxyDispatchAdapter$DispatchingInvocati
> >>>>>>>>>>>> on
> >>>>>>>>>>>> Ha
> >>>>>>>>>>>>>>> ndler.invoke(ProxyDispatchAdapter.java:93)
> >>>>>>>>>>>>>>>    at com.sun.proxy.$Proxy46.executionFinished(Unknown
> >>>>>>>>>>> Source)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.process.internal.DefaultExecHandle.setEndStateInfo(DefaultE
> >>>>>>>>>>>> xe
> >>>>>>>>>>>> cH
> >>>>>>>>>>>>>>> andle.java:212)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.process.internal.DefaultExecHandle.finished(DefaultExecHand
> >>>>>>>>>>>> le
> >>>>>>>>>>>> .j
> >>>>>>>>>>>>>>> ava:309)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.process.internal.ExecHandleRunner.completed(ExecHandleRunne
> >>>>>>>>>>>> r.
> >>>>>>>>>>>> ja
> >>>>>>>>>>>>>>> va:108)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.process.internal.ExecHandleRunner.run(ExecHandleRunner.java
> >>>>>>>>>>>> :8
> >>>>>>>>>>>> 8)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecuto
> >>>>>>>>>>>> rI
> >>>>>>>>>>>> mp
> >>>>>>>>>>>>>>> l$1.run(DefaultExecutorFactory.java:66)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j
> >>>>>>>>>>>> av
> >>>>>>>>>>>> a:
> >>>>>>>>>>>>>>> 1145)
> >>>>>>>>>>>>>>>    at
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>
> >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.
> >>>>>>>>>>>> ja
> >>>>>>>>>>>> va
> >>>>>>>>>>>>>>> :615)
> >>>>>>>>>>>>>>>    at java.lang.Thread.r
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>> Do I need more memory for my machines? Each already has
> >>>>>> 4GB. I
> >>>>>>>>>>> really
> >>>>>>>>>>>>>>> need to have this running. I¹m not sure which way is best
> >>>>>>>>> http or
> >>>>>>>>>>> hdfs
> >>>>>>>>>>>>>>> which one you suggest and how can i solve my problem for
> >>>>>> each
> >>>>>>>>>>> case.
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>> Thanks in advance and sorry for bothering this much.
> >>>>>>>>>>>>>>> On 10 Aug 2014, at 00:20, Telles Nobrega
> >>>>>>>>>>> <[email protected]>
> >>>>>>>>>>>>> wrote:
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>> Hi Chris, now I have the tar file in my RM machine, and
> >>>>>> the
> >>>>>>>>>>> yarn
> >>>>>>>>>>>> path
> >>>>>>>>>>>>>>>> points to it. I changed the core-site.xml to use
> >>>>>>>>> HttpFileSystem
> >>>>>>>>>>>> instead
> >>>>>>>>>>>>>>>> of HDFS now it is failing with
> >>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>> Application application_1407640485281_0001 failed 2
> >>>>> times
> >>>>>>>>> due
> >>>>>>>>>>> to
> >>>>>>>>>>> AM
> >>>>>>>>>>>>>>>> Container for appattempt_1407640485281_0001_000002 exited
> >>>>>>>>> with
> >>>>>>>>>>>>>>>> exitCode:-1000 due to: java.lang.ClassNotFoundException:
> >>>>>>>>> Class
> >>>>>>>>>>>>>>>> org.apache.samza.util.hadoop.HttpFileSystem not found
> >>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>> I think I can solve this just installing scala files
> >>>>> from
> >>>>>>>>> the
> >>>>>>>>>>> samza
> >>>>>>>>>>>>>>>> tutorial, can you confirm that?
> >>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>> On 09 Aug 2014, at 08:34, Telles Nobrega
> >>>>>>>>>>> <[email protected]
> >>>>>>>>>>>>
> >>>>>>>>>>>>>>>> wrote:
> >>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>> Hi Chris,
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>> I think the problem is that I forgot to update the
> >>>>>>>>>>>> yarn.job.package.
> >>>>>>>>>>>>>>>>> I will try again to see if it works now.
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>> I have one more question, how can I stop (command line)
> >>>>>> the
> >>>>>>>>>>> jobs
> >>>>>>>>>>>>>>>>> running in my topology, for the experiment that I will
> >>>>>> run,
> >>>>>>>>> I
> >>>>>>>>>>> need
> >>>>>>>>>>>> to
> >>>>>>>>>>>>>>>>> run the same job in 4 minutes intervals. So I need to
> >>>>> kill
> >>>>>>>>> it,
> >>>>>>>>>>> clean
> >>>>>>>>>>>>>>>>> the kafka topics and rerun.
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>> Thanks in advance.
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>> On 08 Aug 2014, at 12:41, Chris Riccomini
> >>>>>>>>>>>>>>>>> <[email protected]> wrote:
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>> Hey Telles,
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>> Do I need to have the job folder on each machine in
> >>>>> my
> >>>>>>>>>>> cluster?
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>> No, you should not need to do this. There are two ways
> >>>>>> to
> >>>>>>>>>>> deploy
> >>>>>>>>>>>>> your
> >>>>>>>>>>>>>>>>>> tarball to the YARN grid. One is to put it in HDFS,
> >>>>> and
> >>>>>>>>> the
> >>>>>>>>>>> other
> >>>>>>>>>>>> is
> >>>>>>>>>>>>>>>>>> to
> >>>>>>>>>>>>>>>>>> put it on an HTTP server. The link to running a Samza
> >>>>>> job
> >>>>>>>>> in
> >>>>>>>>>>> a
> >>>>>>>>>>>>>>>>>> multi-node
> >>>>>>>>>>>>>>>>>> YARN cluster describes how to do both (either HTTP
> >>>>>> server
> >>>>>>>>> or
> >>>>>>>>>>>> HDFS).
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>> In both cases, once the tarball is put in on the
> >>>>>> HTTP/HDFS
> >>>>>>>>>>>>> server(s),
> >>>>>>>>>>>>>>>>>> you
> >>>>>>>>>>>>>>>>>> must update yarn.package.path to point to it. From
> >>>>>> there,
> >>>>>>>>> the
> >>>>>>>>>>> YARN
> >>>>>>>>>>>>> NM
> >>>>>>>>>>>>>>>>>> should download it for you automatically when you
> >>>>> start
> >>>>>>>>> your
> >>>>>>>>>>> job.
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>> * Can you send along a paste of your job config?
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>> Cheers,
> >>>>>>>>>>>>>>>>>> Chris
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>> On 8/8/14 8:04 AM, "Claudio Martins"
> >>>>>>>>>>> <[email protected]>
> >>>>>>>>>>>>> wrote:
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>> Hi Telles, it looks to me that you forgot to update
> >>>>> the
> >>>>>>>>>>>>>>>>>>> "yarn.package.path"
> >>>>>>>>>>>>>>>>>>> attribute in your config file for the task.
> >>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>> - Claudio Martins
> >>>>>>>>>>>>>>>>>>> Head of Engineering
> >>>>>>>>>>>>>>>>>>> MobileAware USA Inc. / www.mobileaware.com
> >>>>>>>>>>>>>>>>>>> office: +1 617 986 5060 / mobile: +1 617 480 5288
> >>>>>>>>>>>>>>>>>>> linkedin: www.linkedin.com/in/martinsclaudio
> >>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>> On Fri, Aug 8, 2014 at 10:55 AM, Telles Nobrega
> >>>>>>>>>>>>>>>>>>> <[email protected]>
> >>>>>>>>>>>>>>>>>>> wrote:
> >>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>> Hi,
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>> this is my first time trying to run a job on a
> >>>>>> multinode
> >>>>>>>>>>>>>>>>>>>> environment. I
> >>>>>>>>>>>>>>>>>>>> have the cluster set up, I can see in the GUI that
> >>>>> all
> >>>>>>>>>>> nodes
> >>>>>>>>>>> are
> >>>>>>>>>>>>>>>>>>>> working.
> >>>>>>>>>>>>>>>>>>>> Do I need to have the job folder on each machine in
> >>>>> my
> >>>>>>>>>>> cluster?
> >>>>>>>>>>>>>>>>>>>> - The first time I tried running with the job on the
> >>>>>>>>>>> namenode
> >>>>>>>>>>>>>>>>>>>> machine
> >>>>>>>>>>>>>>>>>>>> and
> >>>>>>>>>>>>>>>>>>>> it failed saying:
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>> Application application_1407509228798_0001 failed 2
> >>>>>>>>> times
> >>>>>>>>>>> due
> >>>>>>>>>>> to
> >>>>>>>>>>>>> AM
> >>>>>>>>>>>>>>>>>>>> Container for appattempt_1407509228798_0001_000002
> >>>>>>>>> exited
> >>>>>>>>>>> with
> >>>>>>>>>>>>>>>>>>>> exitCode:
> >>>>>>>>>>>>>>>>>>>> -1000 due to: File
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>>>>>>
> >> file:/home/ubuntu/alarm-samza/samza-job-package/target/samza-job-
> >>>>>>>>>>>>>>>>> pa
> >>>>>>>>>>>>>>>>> ck
> >>>>>>>>>>>>>>>>>>>> age-
> >>>>>>>>>>>>>>>>>>>> 0.7.0-dist.tar.gz
> >>>>>>>>>>>>>>>>>>>> does not exist
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>> So I copied the folder to each machine in my cluster
> >>>>>> and
> >>>>>>>>>>> got
> >>>>>>>>>>>> this
> >>>>>>>>>>>>>>>>>>>> error:
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>> Application application_1407509228798_0002 failed 2
> >>>>>>>>> times
> >>>>>>>>>>> due
> >>>>>>>>>>> to
> >>>>>>>>>>>>> AM
> >>>>>>>>>>>>>>>>>>>> Container for appattempt_1407509228798_0002_000002
> >>>>>>>>> exited
> >>>>>>>>>>> with
> >>>>>>>>>>>>>>>>>>>> exitCode:
> >>>>>>>>>>>>>>>>>>>> -1000 due to: Resource
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>>>>>>>>>>>
> >> file:/home/ubuntu/alarm-samza/samza-job-package/target/samza-job-
> >>>>>>>>>>>>>>>>> pa
> >>>>>>>>>>>>>>>>> ck
> >>>>>>>>>>>>>>>>>>>> age-
> >>>>>>>>>>>>>>>>>>>> 0.7.0-dist.tar.gz
> >>>>>>>>>>>>>>>>>>>> changed on src filesystem (expected 1407509168000,
> >>>>> was
> >>>>>>>>>>>>> 1407509434000
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>> What am I missing?
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>> p.s.: I followed this
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>> <
> >>>>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>
> https://github.com/yahoo/samoa/wiki/Executing-SAMOA-with-Apache-Samz
> >>>>>>>>>>>>>>>>>>>> a>
> >>>>>>>>>>>>>>>>>>>> tutorial
> >>>>>>>>>>>>>>>>>>>> and this
> >>>>>>>>>>>>>>>>>>>> <
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>
> >> http://samza.incubator.apache.org/learn/tutorials/0.7.0/run-in-multi-
> >>>>>>>>>>>>>>>>>>>> node
> >>>>>>>>>>>>>>>>>>>> -yarn.html
> >>>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>> to
> >>>>>>>>>>>>>>>>>>>> set up the cluster.
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>> Help is much appreciated.
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>> Thanks in advance.
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>>> --
> >>>>>>>>>>>>>>>>>>>> ------------------------------------------
> >>>>>>>>>>>>>>>>>>>> Telles Mota Vidal Nobrega
> >>>>>>>>>>>>>>>>>>>> M.sc. Candidate at UFCG
> >>>>>>>>>>>>>>>>>>>> B.sc. in Computer Science at UFCG
> >>>>>>>>>>>>>>>>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG
> >>>>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>>
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>> --
> >>>>>>>>>>>> ------------------------------------------
> >>>>>>>>>>>> Telles Mota Vidal Nobrega
> >>>>>>>>>>>> M.sc. Candidate at UFCG
> >>>>>>>>>>>> B.sc. in Computer Science at UFCG
> >>>>>>>>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG
> >>>>>>>>>>>>
> >>>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>>
> >>>>>>>>>> --
> >>>>>>>>>> ------------------------------------------
> >>>>>>>>>> Telles Mota Vidal Nobrega
> >>>>>>>>>> M.sc. Candidate at UFCG
> >>>>>>>>>> B.sc. in Computer Science at UFCG
> >>>>>>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG
> >>>>>>>>>
> >>>>>>>>>
> >>>>>>>>
> >>>>>>>>
> >>>>>>>> --
> >>>>>>>> ------------------------------------------
> >>>>>>>> Telles Mota Vidal Nobrega
> >>>>>>>> M.sc. Candidate at UFCG
> >>>>>>>> B.sc. in Computer Science at UFCG
> >>>>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG
> >>>>>>>
> >>>>>>>
> >>>>>>
> >>>>>
> >>>>>
> >>>>>
> >>>>> --
> >>>>> ------------------------------------------
> >>>>> Telles Mota Vidal Nobrega
> >>>>> M.sc. Candidate at UFCG
> >>>>> B.sc. in Computer Science at UFCG
> >>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG
> >>>>>
> >>>
> >>
> >>
>
>

Reply via email to