Hi, I copied hadoop-hdfs-2.3.0 to my-job/lib and it changed the error which is 
good but the error is back to 

Exception in thread "main" java.lang.RuntimeException: 
java.lang.ClassNotFoundException: Class 
org.apache.hadoop.hdfs.DistributedFileSystem not found
        at 
org.apache.hadoop.conf.Configuration.getClass(Configuration.java:1720)
        at 
org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2415)
        at 
org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2428)
        at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88)
        at 
org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467)
        at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449)
        at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367)
        at org.apache.hadoop.fs.Path.getFileSystem(Path.java:287)
        at 
org.apache.samza.job.yarn.ClientHelper.submitApplication(ClientHelper.scala:111)
        at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:55)
        at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:48)
        at org.apache.samza.job.JobRunner.run(JobRunner.scala:62)
        at org.apache.samza.job.JobRunner$.main(JobRunner.scala:37)
        at org.apache.samza.job.JobRunner.main(JobRunner.scala)
Caused by: java.lang.ClassNotFoundException: Class 
org.apache.hadoop.hdfs.DistributedFileSystem not found
        at 
org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:1626)
        at 
org.apache.hadoop.conf.Configuration.getClass(Configuration.java:1718)
        ... 13 more

Do I need to have this lib in all nodes at the job folder or just to submit?

On 11 Aug 2014, at 20:11, Yan Fang <[email protected]> wrote:

> Hi Telles,
> 
> I replayed your problem and think I figured out why CLASSPATH does not
> work. Because in our script bin/run-class.sh, we have the line
> "CLASSPATH=$HADOOP_CONF_DIR", which actually ingores your setting.
> 
> So a simple solution is to copy the hadoop-hdfs.jar to your samza lib
> directory. Then run bin/run-job ----config-factory=... --config-path=... .
> Let me know how it goes. Thank you.
> 
> Cheers,
> 
> Fang, Yan
> [email protected]
> +1 (206) 849-4108
> 
> 
> On Mon, Aug 11, 2014 at 4:07 PM, Telles Nobrega <[email protected]>
> wrote:
> 
>> Sure, thanks.
>> 
>> 
>> On Mon, Aug 11, 2014 at 6:22 PM, Yan Fang <[email protected]> wrote:
>> 
>>> Hi Telles,
>>> 
>>> I am not sure whether exporting the CLASSPATH works. (sometimes it does
>> not
>>> work for me...) My suggestion is to include the hdfs jar explicitly in
>> the
>>> package that you upload to hdfs. Also , remember to put the jar into your
>>> local samza (which is deploy/samza/lib if you go with the hello-samza
>>> tutorial) Let me know if that works.
>>> 
>>> Cheers,
>>> 
>>> Fang, Yan
>>> [email protected]
>>> +1 (206) 849-4108
>>> 
>>> 
>>> On Mon, Aug 11, 2014 at 2:04 PM, Chris Riccomini <
>>> [email protected]> wrote:
>>> 
>>>> Hey Telles,
>>>> 
>>>> Hmm. I'm out of ideas. If Zhijie is around, he'd probably be of use,
>> but
>>> I
>>>> haven't heard from him in a while.
>>>> 
>>>> I'm afraid your best bet is probably to email the YARN dev mailing
>> list,
>>>> since this is a YARN config issue.
>>>> 
>>>> Cheers,
>>>> Chris
>>>> 
>>>> On 8/11/14 1:58 PM, "Telles Nobrega" <[email protected]> wrote:
>>>> 
>>>>> ​I exported ​export
>>>> 
>>> 
>>> CLASSPATH=$CLASSPATH:hadoop-2.3.0/share/hadoop/hdfs/hadoop-hdfs-2.3.0.jar
>>>>> and still happened the same problem.
>>>>> 
>>>>> 
>>>>> On Mon, Aug 11, 2014 at 5:35 PM, Chris Riccomini <
>>>>> [email protected]> wrote:
>>>>> 
>>>>>> Hey Telles,
>>>>>> 
>>>>>> It sounds like either the HDFS jar is missing from the classpath, or
>>> the
>>>>>> hdfs file system needs to be configured:
>>>>>> 
>>>>>> <property>
>>>>>>  <name>fs.hdfs.impl</name>
>>>>>>  <value>org.apache.hadoop.hdfs.DistributedFileSystem</value>
>>>>>>  <description>The FileSystem for hdfs: uris.</description>
>>>>>> </property>
>>>>>> 
>>>>>> 
>>>>>> (from
>>>>>> 
>>>>>> 
>>>> 
>>> 
>> https://groups.google.com/a/cloudera.org/forum/#!topic/scm-users/lyho8ptA
>>>>>> zE
>>>>>> 0)
>>>>>> 
>>>>>> I believe this will need to be configured for your NM.
>>>>>> 
>>>>>> Cheers,
>>>>>> Chris
>>>>>> 
>>>>>> On 8/11/14 1:31 PM, "Telles Nobrega" <[email protected]>
>> wrote:
>>>>>> 
>>>>>>> Yes, it is like this:
>>>>>>> 
>>>>>>> <configuration>
>>>>>>> <property>
>>>>>>>   <name>dfs.datanode.data.dir</name>
>>>>>>>   <value>file:///home/ubuntu/hadoop-2.3.0/hdfs/datanode</value>
>>>>>>>   <description>Comma separated list of paths on the local
>>> filesystem
>>>>>> of
>>>>>>> a
>>>>>>> DataNode where it should store its blocks.</description>
>>>>>>> </property>
>>>>>>> 
>>>>>>> <property>
>>>>>>>   <name>dfs.namenode.name.dir</name>
>>>>>>>   <value>file:///home/ubuntu/hadoop-2.3.0/hdfs/namenode</value>
>>>>>>>   <description>Path on the local filesystem where the NameNode
>>> stores
>>>>>>> the
>>>>>>> namespace and transaction logs persistently.</description>
>>>>>>> </property>
>>>>>>> </configuration>
>>>>>>> ~
>>>>>>> 
>>>>>>> I saw some report that this may be a classpath problem. Does this
>>>>>> sounds
>>>>>>> right to you?
>>>>>>> 
>>>>>>> 
>>>>>>> On Mon, Aug 11, 2014 at 5:25 PM, Yan Fang <[email protected]>
>>>> wrote:
>>>>>>> 
>>>>>>>> Hi Telles,
>>>>>>>> 
>>>>>>>> It looks correct. Did you put the hdfs-site.xml into your
>>>>>>>> HADOOP_CONF_DIR
>>>>>>>> ?(such as ~/.samza/conf)
>>>>>>>> 
>>>>>>>> Fang, Yan
>>>>>>>> [email protected]
>>>>>>>> +1 (206) 849-4108
>>>>>>>> 
>>>>>>>> 
>>>>>>>> On Mon, Aug 11, 2014 at 1:02 PM, Telles Nobrega
>>>>>>>> <[email protected]>
>>>>>>>> wrote:
>>>>>>>> 
>>>>>>>>> ​Hi Yan Fang,
>>>>>>>>> 
>>>>>>>>> I was able to deploy the file to hdfs, I can see them in all my
>>>>>> nodes
>>>>>>>> but
>>>>>>>>> when I tried running I got this error:
>>>>>>>>> 
>>>>>>>>> Exception in thread "main" java.io.IOException: No FileSystem
>> for
>>>>>>>> scheme:
>>>>>>>>> hdfs
>>>>>>>>> at
>>>>>>>> 
>>>> 
>>> 
>>>> org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2421)
>>>>>>>>> at
>>>>>>>> 
>>>> 
>>>> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2428)
>>>>>>>>> at
>> org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88)
>>>>>>>>> at
>>>>>>>> 
>>>> 
>>>> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467)
>>>>>>>>> at
>>> org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449)
>>>>>>>>> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367)
>>>>>>>>> at org.apache.hadoop.fs.Path.getFileSystem(Path.java:287)
>>>>>>>>> at
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>> org.apache.samza.job.yarn.ClientHelper.submitApplication(ClientHelper.s
>>>>>>>> ca
>>>>>>>> la:111)
>>>>>>>>> at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:55)
>>>>>>>>> at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:48)
>>>>>>>>> at org.apache.samza.job.JobRunner.run(JobRunner.scala:62)
>>>>>>>>> at org.apache.samza.job.JobRunner$.main(JobRunner.scala:37)
>>>>>>>>> at org.apache.samza.job.JobRunner.main(JobRunner.scala)
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> This is my yarn.package.path config:
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>> ​yarn.package.path=hdfs://telles-master-samza:50070/samza-job-package-0
>>>>>>>> .7
>>>>>>>> .0-dist.tar.gz
>>>>>>>>> 
>>>>>>>>> Thanks in advance
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> On Mon, Aug 11, 2014 at 3:00 PM, Yan Fang <
>> [email protected]>
>>>>>>>> wrote:
>>>>>>>>> 
>>>>>>>>>> Hi Telles,
>>>>>>>>>> 
>>>>>>>>>> In terms of "*I tried pushing the tar file to HDFS but I got
>> an
>>>>>>>> error
>>>>>>>>> from
>>>>>>>>>> hadoop saying that it couldn’t find core-site.xml file*.", I
>>>>>> guess
>>>>>>>> you
>>>>>>>>> set
>>>>>>>>>> the HADOOP_CONF_DIR variable and made it point to
>>> ~/.samza/conf.
>>>>>> You
>>>>>>>> can
>>>>>>>>> do
>>>>>>>>>> 1) make the HADOOP_CONF_DIR point to the directory where your
>>>>>> conf
>>>>>>>> files
>>>>>>>>>> are, such as /etc/hadoop/conf. Or 2) copy the config files to
>>>>>>>>>> ~/.samza/conf. Thank you,
>>>>>>>>>> 
>>>>>>>>>> Cheer,
>>>>>>>>>> 
>>>>>>>>>> Fang, Yan
>>>>>>>>>> [email protected]
>>>>>>>>>> +1 (206) 849-4108
>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>>> On Mon, Aug 11, 2014 at 7:40 AM, Chris Riccomini <
>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>> 
>>>>>>>>>>> Hey Telles,
>>>>>>>>>>> 
>>>>>>>>>>> To get YARN working with the HTTP file system, you need to
>>>>>> follow
>>>>>>>> the
>>>>>>>>>>> instructions on:
>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>>>> 
>>>> 
>>> 
>> http://samza.incubator.apache.org/learn/tutorials/0.7.0/run-in-multi-node
>>>>>>>> -y
>>>>>>>>>>> arn.html
>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>>> In the "Set Up Http Filesystem for YARN" section.
>>>>>>>>>>> 
>>>>>>>>>>> You shouldn't need to compile anything (no Gradle, which is
>>>>>> what
>>>>>>>> your
>>>>>>>>>>> stack trace is showing). This setup should be done for all
>> of
>>>>>> the
>>>>>>>> NMs,
>>>>>>>>>>> since they will be the ones downloading your job's package
>>>>>> (from
>>>>>>>>>>> yarn.package.path).
>>>>>>>>>>> 
>>>>>>>>>>> Cheers,
>>>>>>>>>>> Chris
>>>>>>>>>>> 
>>>>>>>>>>> On 8/9/14 9:44 PM, "Telles Nobrega" <
>> [email protected]
>>>> 
>>>>>>>> wrote:
>>>>>>>>>>> 
>>>>>>>>>>>> Hi again, I tried installing the scala libs but the Http
>>>>>> problem
>>>>>>>> still
>>>>>>>>>>>> occurs. I realised that I need to compile incubator samza
>> in
>>>>>> the
>>>>>>>>>> machines
>>>>>>>>>>>> that I¹m going to run the jobs, but the compilation fails
>>> with
>>>>>>>> this
>>>>>>>>> huge
>>>>>>>>>>>> message:
>>>>>>>>>>>> 
>>>>>>>>>>>> #
>>>>>>>>>>>> # There is insufficient memory for the Java Runtime
>>>>>> Environment
>>>>>>>> to
>>>>>>>>>>>> continue.
>>>>>>>>>>>> # Native memory allocation (malloc) failed to allocate
>>>>>> 3946053632
>>>>>>>>> bytes
>>>>>>>>>>>> for committing reserved memory.
>>>>>>>>>>>> # An error report file with more information is saved as:
>>>>>>>>>>>> #
>>> /home/ubuntu/incubator-samza/samza-kafka/hs_err_pid2506.log
>>>>>>>>>>>> Could not write standard input into: Gradle Worker 13.
>>>>>>>>>>>> java.io.IOException: Broken pipe
>>>>>>>>>>>>      at java.io.FileOutputStream.writeBytes(Native
>> Method)
>>>>>>>>>>>>      at
>>>>>>>> java.io.FileOutputStream.write(FileOutputStream.java:345)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>> 
>>>> 
>>>> java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>> java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.process.internal.streams.ExecOutputHandleRunner.run(ExecOut
>>>>>>>>> pu
>>>>>>>>> tH
>>>>>>>>>>>> andleRunner.java:53)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecuto
>>>>>>>>> rI
>>>>>>>>> mp
>>>>>>>>>>>> l$1.run(DefaultExecutorFactory.java:66)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j
>>>>>>>>> av
>>>>>>>>> a:
>>>>>>>>>>>> 1145)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.
>>>>>>>>> ja
>>>>>>>>> va
>>>>>>>>>>>> :615)
>>>>>>>>>>>>      at java.lang.Thread.run(Thread.java:744)
>>>>>>>>>>>> Process 'Gradle Worker 13' finished with non-zero exit
>>> value 1
>>>>>>>>>>>> org.gradle.process.internal.ExecException: Process 'Gradle
>>>>>> Worker
>>>>>>>> 13'
>>>>>>>>>>>> finished with non-zero exit value 1
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.process.internal.DefaultExecHandle$ExecResultImpl.assertNor
>>>>>>>>> ma
>>>>>>>>> lE
>>>>>>>>>>>> xitValue(DefaultExecHandle.java:362)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.process.internal.DefaultWorkerProcess.onProcessStop(Default
>>>>>>>>> Wo
>>>>>>>>> rk
>>>>>>>>>>>> erProcess.java:89)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.process.internal.DefaultWorkerProcess.access$000(DefaultWor
>>>>>>>>> ke
>>>>>>>>> rP
>>>>>>>>>>>> rocess.java:33)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.process.internal.DefaultWorkerProcess$1.executionFinished(D
>>>>>>>>> ef
>>>>>>>>> au
>>>>>>>>>>>> ltWorkerProcess.java:55)
>>>>>>>>>>>>      at
>>> sun.reflect.NativeMethodAccessorImpl.invoke0(Native
>>>>>>>> Method)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j
>>>>>>>>> av
>>>>>>>>> a:
>>>>>>>>>>>> 57)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccess
>>>>>>>>> or
>>>>>>>>> Im
>>>>>>>>>>>> pl.java:43)
>>>>>>>>>>>>      at java.lang.reflect.Method.invoke(Method.java:606)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDi
>>>>>>>>> sp
>>>>>>>>> at
>>>>>>>>>>>> ch.java:35)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDi
>>>>>>>>> sp
>>>>>>>>> at
>>>>>>>>>>>> ch.java:24)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java:
>>>>>>>>> 81
>>>>>>>>> )
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java:
>>>>>>>>> 30
>>>>>>>>> )
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.messaging.dispatch.ProxyDispatchAdapter$DispatchingInvocati
>>>>>>>>> on
>>>>>>>>> Ha
>>>>>>>>>>>> ndler.invoke(ProxyDispatchAdapter.java:93)
>>>>>>>>>>>>      at com.sun.proxy.$Proxy46.executionFinished(Unknown
>>>>>>>> Source)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.process.internal.DefaultExecHandle.setEndStateInfo(DefaultE
>>>>>>>>> xe
>>>>>>>>> cH
>>>>>>>>>>>> andle.java:212)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.process.internal.DefaultExecHandle.finished(DefaultExecHand
>>>>>>>>> le
>>>>>>>>> .j
>>>>>>>>>>>> ava:309)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.process.internal.ExecHandleRunner.completed(ExecHandleRunne
>>>>>>>>> r.
>>>>>>>>> ja
>>>>>>>>>>>> va:108)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.process.internal.ExecHandleRunner.run(ExecHandleRunner.java
>>>>>>>>> :8
>>>>>>>>> 8)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecuto
>>>>>>>>> rI
>>>>>>>>> mp
>>>>>>>>>>>> l$1.run(DefaultExecutorFactory.java:66)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j
>>>>>>>>> av
>>>>>>>>> a:
>>>>>>>>>>>> 1145)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.
>>>>>>>>> ja
>>>>>>>>> va
>>>>>>>>>>>> :615)
>>>>>>>>>>>>      at java.lang.Thread.run(Thread.java:744)
>>>>>>>>>>>> OpenJDK 64-Bit Server VM warning: INFO:
>>>>>>>>>>>> os::commit_memory(0x000000070a6c0000, 3946053632, 0)
>> failed;
>>>>>>>>>>>> error='Cannot allocate memory' (errno=12)
>>>>>>>>>>>> #
>>>>>>>>>>>> # There is insufficient memory for the Java Runtime
>>>>>> Environment
>>>>>>>> to
>>>>>>>>>>>> continue.
>>>>>>>>>>>> # Native memory allocation (malloc) failed to allocate
>>>>>> 3946053632
>>>>>>>>> bytes
>>>>>>>>>>>> for committing reserved memory.
>>>>>>>>>>>> # An error report file with more information is saved as:
>>>>>>>>>>>> #
>>> /home/ubuntu/incubator-samza/samza-kafka/hs_err_pid2518.log
>>>>>>>>>>>> Could not write standard input into: Gradle Worker 14.
>>>>>>>>>>>> java.io.IOException: Broken pipe
>>>>>>>>>>>>      at java.io.FileOutputStream.writeBytes(Native
>> Method)
>>>>>>>>>>>>      at
>>>>>>>> java.io.FileOutputStream.write(FileOutputStream.java:345)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>> 
>>>> 
>>>> java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>> java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.process.internal.streams.ExecOutputHandleRunner.run(ExecOut
>>>>>>>>> pu
>>>>>>>>> tH
>>>>>>>>>>>> andleRunner.java:53)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecuto
>>>>>>>>> rI
>>>>>>>>> mp
>>>>>>>>>>>> l$1.run(DefaultExecutorFactory.java:66)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j
>>>>>>>>> av
>>>>>>>>> a:
>>>>>>>>>>>> 1145)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.
>>>>>>>>> ja
>>>>>>>>> va
>>>>>>>>>>>> :615)
>>>>>>>>>>>>      at java.lang.Thread.run(Thread.java:744)
>>>>>>>>>>>> Process 'Gradle Worker 14' finished with non-zero exit
>>> value 1
>>>>>>>>>>>> org.gradle.process.internal.ExecException: Process 'Gradle
>>>>>> Worker
>>>>>>>> 14'
>>>>>>>>>>>> finished with non-zero exit value 1
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.process.internal.DefaultExecHandle$ExecResultImpl.assertNor
>>>>>>>>> ma
>>>>>>>>> lE
>>>>>>>>>>>> xitValue(DefaultExecHandle.java:362)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.process.internal.DefaultWorkerProcess.onProcessStop(Default
>>>>>>>>> Wo
>>>>>>>>> rk
>>>>>>>>>>>> erProcess.java:89)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.process.internal.DefaultWorkerProcess.access$000(DefaultWor
>>>>>>>>> ke
>>>>>>>>> rP
>>>>>>>>>>>> rocess.java:33)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.process.internal.DefaultWorkerProcess$1.executionFinished(D
>>>>>>>>> ef
>>>>>>>>> au
>>>>>>>>>>>> ltWorkerProcess.java:55)
>>>>>>>>>>>>      at
>>> sun.reflect.NativeMethodAccessorImpl.invoke0(Native
>>>>>>>> Method)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j
>>>>>>>>> av
>>>>>>>>> a:
>>>>>>>>>>>> 57)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccess
>>>>>>>>> or
>>>>>>>>> Im
>>>>>>>>>>>> pl.java:43)
>>>>>>>>>>>>      at java.lang.reflect.Method.invoke(Method.java:606)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDi
>>>>>>>>> sp
>>>>>>>>> at
>>>>>>>>>>>> ch.java:35)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDi
>>>>>>>>> sp
>>>>>>>>> at
>>>>>>>>>>>> ch.java:24)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java:
>>>>>>>>> 81
>>>>>>>>> )
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java:
>>>>>>>>> 30
>>>>>>>>> )
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.messaging.dispatch.ProxyDispatchAdapter$DispatchingInvocati
>>>>>>>>> on
>>>>>>>>> Ha
>>>>>>>>>>>> ndler.invoke(ProxyDispatchAdapter.java:93)
>>>>>>>>>>>>      at com.sun.proxy.$Proxy46.executionFinished(Unknown
>>>>>>>> Source)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.process.internal.DefaultExecHandle.setEndStateInfo(DefaultE
>>>>>>>>> xe
>>>>>>>>> cH
>>>>>>>>>>>> andle.java:212)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.process.internal.DefaultExecHandle.finished(DefaultExecHand
>>>>>>>>> le
>>>>>>>>> .j
>>>>>>>>>>>> ava:309)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.process.internal.ExecHandleRunner.completed(ExecHandleRunne
>>>>>>>>> r.
>>>>>>>>> ja
>>>>>>>>>>>> va:108)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.process.internal.ExecHandleRunner.run(ExecHandleRunner.java
>>>>>>>>> :8
>>>>>>>>> 8)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecuto
>>>>>>>>> rI
>>>>>>>>> mp
>>>>>>>>>>>> l$1.run(DefaultExecutorFactory.java:66)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j
>>>>>>>>> av
>>>>>>>>> a:
>>>>>>>>>>>> 1145)
>>>>>>>>>>>>      at
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.
>>>>>>>>> ja
>>>>>>>>> va
>>>>>>>>>>>> :615)
>>>>>>>>>>>>      at java.lang.Thread.r
>>>>>>>>>>>> 
>>>>>>>>>>>> Do I need more memory for my machines? Each already has
>>> 4GB. I
>>>>>>>> really
>>>>>>>>>>>> need to have this running. I¹m not sure which way is best
>>>>>> http or
>>>>>>>> hdfs
>>>>>>>>>>>> which one you suggest and how can i solve my problem for
>>> each
>>>>>>>> case.
>>>>>>>>>>>> 
>>>>>>>>>>>> Thanks in advance and sorry for bothering this much.
>>>>>>>>>>>> On 10 Aug 2014, at 00:20, Telles Nobrega
>>>>>>>> <[email protected]>
>>>>>>>>>> wrote:
>>>>>>>>>>>> 
>>>>>>>>>>>>> Hi Chris, now I have the tar file in my RM machine, and
>>> the
>>>>>>>> yarn
>>>>>>>>> path
>>>>>>>>>>>>> points to it. I changed the core-site.xml to use
>>>>>> HttpFileSystem
>>>>>>>>> instead
>>>>>>>>>>>>> of HDFS now it is failing with
>>>>>>>>>>>>> 
>>>>>>>>>>>>> Application application_1407640485281_0001 failed 2
>> times
>>>>>> due
>>>>>>>> to
>>>>>>>> AM
>>>>>>>>>>>>> Container for appattempt_1407640485281_0001_000002 exited
>>>>>> with
>>>>>>>>>>>>> exitCode:-1000 due to: java.lang.ClassNotFoundException:
>>>>>> Class
>>>>>>>>>>>>> org.apache.samza.util.hadoop.HttpFileSystem not found
>>>>>>>>>>>>> 
>>>>>>>>>>>>> I think I can solve this just installing scala files
>> from
>>>>>> the
>>>>>>>> samza
>>>>>>>>>>>>> tutorial, can you confirm that?
>>>>>>>>>>>>> 
>>>>>>>>>>>>> On 09 Aug 2014, at 08:34, Telles Nobrega
>>>>>>>> <[email protected]
>>>>>>>>> 
>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Hi Chris,
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> I think the problem is that I forgot to update the
>>>>>>>>> yarn.job.package.
>>>>>>>>>>>>>> I will try again to see if it works now.
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> I have one more question, how can I stop (command line)
>>> the
>>>>>>>> jobs
>>>>>>>>>>>>>> running in my topology, for the experiment that I will
>>> run,
>>>>>> I
>>>>>>>> need
>>>>>>>>> to
>>>>>>>>>>>>>> run the same job in 4 minutes intervals. So I need to
>> kill
>>>>>> it,
>>>>>>>> clean
>>>>>>>>>>>>>> the kafka topics and rerun.
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Thanks in advance.
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> On 08 Aug 2014, at 12:41, Chris Riccomini
>>>>>>>>>>>>>> <[email protected]> wrote:
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Hey Telles,
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Do I need to have the job folder on each machine in
>> my
>>>>>>>> cluster?
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> No, you should not need to do this. There are two ways
>>> to
>>>>>>>> deploy
>>>>>>>>>> your
>>>>>>>>>>>>>>> tarball to the YARN grid. One is to put it in HDFS,
>> and
>>>>>> the
>>>>>>>> other
>>>>>>>>> is
>>>>>>>>>>>>>>> to
>>>>>>>>>>>>>>> put it on an HTTP server. The link to running a Samza
>>> job
>>>>>> in
>>>>>>>> a
>>>>>>>>>>>>>>> multi-node
>>>>>>>>>>>>>>> YARN cluster describes how to do both (either HTTP
>>> server
>>>>>> or
>>>>>>>>> HDFS).
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> In both cases, once the tarball is put in on the
>>> HTTP/HDFS
>>>>>>>>>> server(s),
>>>>>>>>>>>>>>> you
>>>>>>>>>>>>>>> must update yarn.package.path to point to it. From
>>> there,
>>>>>> the
>>>>>>>> YARN
>>>>>>>>>> NM
>>>>>>>>>>>>>>> should download it for you automatically when you
>> start
>>>>>> your
>>>>>>>> job.
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> * Can you send along a paste of your job config?
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Cheers,
>>>>>>>>>>>>>>> Chris
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> On 8/8/14 8:04 AM, "Claudio Martins"
>>>>>>>> <[email protected]>
>>>>>>>>>> wrote:
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> Hi Telles, it looks to me that you forgot to update
>> the
>>>>>>>>>>>>>>>> "yarn.package.path"
>>>>>>>>>>>>>>>> attribute in your config file for the task.
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> - Claudio Martins
>>>>>>>>>>>>>>>> Head of Engineering
>>>>>>>>>>>>>>>> MobileAware USA Inc. / www.mobileaware.com
>>>>>>>>>>>>>>>> office: +1 617 986 5060 / mobile: +1 617 480 5288
>>>>>>>>>>>>>>>> linkedin: www.linkedin.com/in/martinsclaudio
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> On Fri, Aug 8, 2014 at 10:55 AM, Telles Nobrega
>>>>>>>>>>>>>>>> <[email protected]>
>>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Hi,
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> this is my first time trying to run a job on a
>>> multinode
>>>>>>>>>>>>>>>>> environment. I
>>>>>>>>>>>>>>>>> have the cluster set up, I can see in the GUI that
>> all
>>>>>>>> nodes
>>>>>>>> are
>>>>>>>>>>>>>>>>> working.
>>>>>>>>>>>>>>>>> Do I need to have the job folder on each machine in
>> my
>>>>>>>> cluster?
>>>>>>>>>>>>>>>>> - The first time I tried running with the job on the
>>>>>>>> namenode
>>>>>>>>>>>>>>>>> machine
>>>>>>>>>>>>>>>>> and
>>>>>>>>>>>>>>>>> it failed saying:
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Application application_1407509228798_0001 failed 2
>>>>>> times
>>>>>>>> due
>>>>>>>> to
>>>>>>>>>> AM
>>>>>>>>>>>>>>>>> Container for appattempt_1407509228798_0001_000002
>>>>>> exited
>>>>>>>> with
>>>>>>>>>>>>>>>>> exitCode:
>>>>>>>>>>>>>>>>> -1000 due to: File
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>>>>>>> file:/home/ubuntu/alarm-samza/samza-job-package/target/samza-job-
>>>>>>>>>>>>>> pa
>>>>>>>>>>>>>> ck
>>>>>>>>>>>>>>>>> age-
>>>>>>>>>>>>>>>>> 0.7.0-dist.tar.gz
>>>>>>>>>>>>>>>>> does not exist
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> So I copied the folder to each machine in my cluster
>>> and
>>>>>>>> got
>>>>>>>>> this
>>>>>>>>>>>>>>>>> error:
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Application application_1407509228798_0002 failed 2
>>>>>> times
>>>>>>>> due
>>>>>>>> to
>>>>>>>>>> AM
>>>>>>>>>>>>>>>>> Container for appattempt_1407509228798_0002_000002
>>>>>> exited
>>>>>>>> with
>>>>>>>>>>>>>>>>> exitCode:
>>>>>>>>>>>>>>>>> -1000 due to: Resource
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>> 
>>>>>> 
>>>> 
>>> 
>>>>>>>>>>>> file:/home/ubuntu/alarm-samza/samza-job-package/target/samza-job-
>>>>>>>>>>>>>> pa
>>>>>>>>>>>>>> ck
>>>>>>>>>>>>>>>>> age-
>>>>>>>>>>>>>>>>> 0.7.0-dist.tar.gz
>>>>>>>>>>>>>>>>> changed on src filesystem (expected 1407509168000,
>> was
>>>>>>>>>> 1407509434000
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> What am I missing?
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> p.s.: I followed this
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> <
>>>>>>>>>>> 
>>>>>>>> 
>>> https://github.com/yahoo/samoa/wiki/Executing-SAMOA-with-Apache-Samz
>>>>>>>>>>>>>>>>> a>
>>>>>>>>>>>>>>>>> tutorial
>>>>>>>>>>>>>>>>> and this
>>>>>>>>>>>>>>>>> <
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>> 
>>>> http://samza.incubator.apache.org/learn/tutorials/0.7.0/run-in-multi-
>>>>>>>>>>>>>>>>> node
>>>>>>>>>>>>>>>>> -yarn.html
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> to
>>>>>>>>>>>>>>>>> set up the cluster.
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Help is much appreciated.
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Thanks in advance.
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> --
>>>>>>>>>>>>>>>>> ------------------------------------------
>>>>>>>>>>>>>>>>> Telles Mota Vidal Nobrega
>>>>>>>>>>>>>>>>> M.sc. Candidate at UFCG
>>>>>>>>>>>>>>>>> B.sc. in Computer Science at UFCG
>>>>>>>>>>>>>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> 
>>>>>>>>>>>>> 
>>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> --
>>>>>>>>> ------------------------------------------
>>>>>>>>> Telles Mota Vidal Nobrega
>>>>>>>>> M.sc. Candidate at UFCG
>>>>>>>>> B.sc. in Computer Science at UFCG
>>>>>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG
>>>>>>>>> 
>>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> --
>>>>>>> ------------------------------------------
>>>>>>> Telles Mota Vidal Nobrega
>>>>>>> M.sc. Candidate at UFCG
>>>>>>> B.sc. in Computer Science at UFCG
>>>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG
>>>>>> 
>>>>>> 
>>>>> 
>>>>> 
>>>>> --
>>>>> ------------------------------------------
>>>>> Telles Mota Vidal Nobrega
>>>>> M.sc. Candidate at UFCG
>>>>> B.sc. in Computer Science at UFCG
>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG
>>>> 
>>>> 
>>> 
>> 
>> 
>> 
>> --
>> ------------------------------------------
>> Telles Mota Vidal Nobrega
>> M.sc. Candidate at UFCG
>> B.sc. in Computer Science at UFCG
>> Software Engineer at OpenStack Project - HP/LSD-UFCG
>> 

Reply via email to