Cool, we are almost there. Could you remove <property> <name>fs.hdfs.impl</name> <value>org.apache.hadoop.hdfs. DistributedFileSystem</value> <description>The FileSystem for hdfs: uris.</description> </property>
To see how it works? Fang, Yan [email protected] +1 (206) 849-4108 On Mon, Aug 11, 2014 at 5:03 PM, Telles Nobrega <[email protected]> wrote: > You may forget this last email, I was really stupid and put the files in a > different folder. Now it could find the file but it’s not there yet… > another error came up > > Exception in thread "main" java.util.ServiceConfigurationError: > org.apache.hadoop.fs.FileSystem: Provider > org.apache.hadoop.hdfs.DistributedFileSystem could not be instantiated > at java.util.ServiceLoader.fail(ServiceLoader.java:224) > at java.util.ServiceLoader.access$100(ServiceLoader.java:181) > at > java.util.ServiceLoader$LazyIterator.next(ServiceLoader.java:377) > at java.util.ServiceLoader$1.next(ServiceLoader.java:445) > at > org.apache.hadoop.fs.FileSystem.loadFileSystems(FileSystem.java:2400) > at > org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2411) > at > org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2428) > at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88) > at > org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467) > at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449) > at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367) > at org.apache.hadoop.fs.Path.getFileSystem(Path.java:287) > at > org.apache.samza.job.yarn.ClientHelper.submitApplication(ClientHelper.scala:111) > at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:55) > at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:48) > at org.apache.samza.job.JobRunner.run(JobRunner.scala:62) > at org.apache.samza.job.JobRunner$.main(JobRunner.scala:37) > at org.apache.samza.job.JobRunner.main(JobRunner.scala) > Caused by: java.lang.NoClassDefFoundError: > org/apache/hadoop/conf/Configuration$DeprecationDelta > at > org.apache.hadoop.hdfs.HdfsConfiguration.addDeprecatedKeys(HdfsConfiguration.java:66) > at > org.apache.hadoop.hdfs.HdfsConfiguration.<clinit>(HdfsConfiguration.java:31) > at > org.apache.hadoop.hdfs.DistributedFileSystem.<clinit>(DistributedFileSystem.java:106) > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native > Method) > at > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) > at > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) > at java.lang.reflect.Constructor.newInstance(Constructor.java:526) > at java.lang.Class.newInstance(Class.java:374) > at > java.util.ServiceLoader$LazyIterator.next(ServiceLoader.java:373) > ... 15 more > Caused by: java.lang.ClassNotFoundException: > org.apache.hadoop.conf.Configuration$DeprecationDelta > at java.net.URLClassLoader$1.run(URLClassLoader.java:366) > at java.net.URLClassLoader$1.run(URLClassLoader.java:355) > at java.security.AccessController.doPrivileged(Native Method) > at java.net.URLClassLoader.findClass(URLClassLoader.java:354) > at java.lang.ClassLoader.loadClass(ClassLoader.java:425) > at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308) > at java.lang.ClassLoader.loadClass(ClassLoader.java:358) > ... 24 more > > On 11 Aug 2014, at 20:45, Telles Nobrega <[email protected]> wrote: > > > Hi, I copied hadoop-hdfs-2.3.0 to my-job/lib and it changed the error > which is good but the error is back to > > > > Exception in thread "main" java.lang.RuntimeException: > java.lang.ClassNotFoundException: Class > org.apache.hadoop.hdfs.DistributedFileSystem not found > > at > org.apache.hadoop.conf.Configuration.getClass(Configuration.java:1720) > > at > org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2415) > > at > org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2428) > > at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88) > > at > org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467) > > at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449) > > at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367) > > at org.apache.hadoop.fs.Path.getFileSystem(Path.java:287) > > at > org.apache.samza.job.yarn.ClientHelper.submitApplication(ClientHelper.scala:111) > > at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:55) > > at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:48) > > at org.apache.samza.job.JobRunner.run(JobRunner.scala:62) > > at org.apache.samza.job.JobRunner$.main(JobRunner.scala:37) > > at org.apache.samza.job.JobRunner.main(JobRunner.scala) > > Caused by: java.lang.ClassNotFoundException: Class > org.apache.hadoop.hdfs.DistributedFileSystem not found > > at > org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:1626) > > at > org.apache.hadoop.conf.Configuration.getClass(Configuration.java:1718) > > ... 13 more > > > > Do I need to have this lib in all nodes at the job folder or just to > submit? > > > > On 11 Aug 2014, at 20:11, Yan Fang <[email protected]> wrote: > > > >> Hi Telles, > >> > >> I replayed your problem and think I figured out why CLASSPATH does not > >> work. Because in our script bin/run-class.sh, we have the line > >> "CLASSPATH=$HADOOP_CONF_DIR", which actually ingores your setting. > >> > >> So a simple solution is to copy the hadoop-hdfs.jar to your samza lib > >> directory. Then run bin/run-job ----config-factory=... > --config-path=... . > >> Let me know how it goes. Thank you. > >> > >> Cheers, > >> > >> Fang, Yan > >> [email protected] > >> +1 (206) 849-4108 > >> > >> > >> On Mon, Aug 11, 2014 at 4:07 PM, Telles Nobrega < > [email protected]> > >> wrote: > >> > >>> Sure, thanks. > >>> > >>> > >>> On Mon, Aug 11, 2014 at 6:22 PM, Yan Fang <[email protected]> > wrote: > >>> > >>>> Hi Telles, > >>>> > >>>> I am not sure whether exporting the CLASSPATH works. (sometimes it > does > >>> not > >>>> work for me...) My suggestion is to include the hdfs jar explicitly in > >>> the > >>>> package that you upload to hdfs. Also , remember to put the jar into > your > >>>> local samza (which is deploy/samza/lib if you go with the hello-samza > >>>> tutorial) Let me know if that works. > >>>> > >>>> Cheers, > >>>> > >>>> Fang, Yan > >>>> [email protected] > >>>> +1 (206) 849-4108 > >>>> > >>>> > >>>> On Mon, Aug 11, 2014 at 2:04 PM, Chris Riccomini < > >>>> [email protected]> wrote: > >>>> > >>>>> Hey Telles, > >>>>> > >>>>> Hmm. I'm out of ideas. If Zhijie is around, he'd probably be of use, > >>> but > >>>> I > >>>>> haven't heard from him in a while. > >>>>> > >>>>> I'm afraid your best bet is probably to email the YARN dev mailing > >>> list, > >>>>> since this is a YARN config issue. > >>>>> > >>>>> Cheers, > >>>>> Chris > >>>>> > >>>>> On 8/11/14 1:58 PM, "Telles Nobrega" <[email protected]> > wrote: > >>>>> > >>>>>> I exported export > >>>>> > >>>> > >>>> > CLASSPATH=$CLASSPATH:hadoop-2.3.0/share/hadoop/hdfs/hadoop-hdfs-2.3.0.jar > >>>>>> and still happened the same problem. > >>>>>> > >>>>>> > >>>>>> On Mon, Aug 11, 2014 at 5:35 PM, Chris Riccomini < > >>>>>> [email protected]> wrote: > >>>>>> > >>>>>>> Hey Telles, > >>>>>>> > >>>>>>> It sounds like either the HDFS jar is missing from the classpath, > or > >>>> the > >>>>>>> hdfs file system needs to be configured: > >>>>>>> > >>>>>>> <property> > >>>>>>> <name>fs.hdfs.impl</name> > >>>>>>> <value>org.apache.hadoop.hdfs.DistributedFileSystem</value> > >>>>>>> <description>The FileSystem for hdfs: uris.</description> > >>>>>>> </property> > >>>>>>> > >>>>>>> > >>>>>>> (from > >>>>>>> > >>>>>>> > >>>>> > >>>> > >>> > https://groups.google.com/a/cloudera.org/forum/#!topic/scm-users/lyho8ptA > >>>>>>> zE > >>>>>>> 0) > >>>>>>> > >>>>>>> I believe this will need to be configured for your NM. > >>>>>>> > >>>>>>> Cheers, > >>>>>>> Chris > >>>>>>> > >>>>>>> On 8/11/14 1:31 PM, "Telles Nobrega" <[email protected]> > >>> wrote: > >>>>>>> > >>>>>>>> Yes, it is like this: > >>>>>>>> > >>>>>>>> <configuration> > >>>>>>>> <property> > >>>>>>>> <name>dfs.datanode.data.dir</name> > >>>>>>>> <value>file:///home/ubuntu/hadoop-2.3.0/hdfs/datanode</value> > >>>>>>>> <description>Comma separated list of paths on the local > >>>> filesystem > >>>>>>> of > >>>>>>>> a > >>>>>>>> DataNode where it should store its blocks.</description> > >>>>>>>> </property> > >>>>>>>> > >>>>>>>> <property> > >>>>>>>> <name>dfs.namenode.name.dir</name> > >>>>>>>> <value>file:///home/ubuntu/hadoop-2.3.0/hdfs/namenode</value> > >>>>>>>> <description>Path on the local filesystem where the NameNode > >>>> stores > >>>>>>>> the > >>>>>>>> namespace and transaction logs persistently.</description> > >>>>>>>> </property> > >>>>>>>> </configuration> > >>>>>>>> ~ > >>>>>>>> > >>>>>>>> I saw some report that this may be a classpath problem. Does this > >>>>>>> sounds > >>>>>>>> right to you? > >>>>>>>> > >>>>>>>> > >>>>>>>> On Mon, Aug 11, 2014 at 5:25 PM, Yan Fang <[email protected]> > >>>>> wrote: > >>>>>>>> > >>>>>>>>> Hi Telles, > >>>>>>>>> > >>>>>>>>> It looks correct. Did you put the hdfs-site.xml into your > >>>>>>>>> HADOOP_CONF_DIR > >>>>>>>>> ?(such as ~/.samza/conf) > >>>>>>>>> > >>>>>>>>> Fang, Yan > >>>>>>>>> [email protected] > >>>>>>>>> +1 (206) 849-4108 > >>>>>>>>> > >>>>>>>>> > >>>>>>>>> On Mon, Aug 11, 2014 at 1:02 PM, Telles Nobrega > >>>>>>>>> <[email protected]> > >>>>>>>>> wrote: > >>>>>>>>> > >>>>>>>>>> Hi Yan Fang, > >>>>>>>>>> > >>>>>>>>>> I was able to deploy the file to hdfs, I can see them in all my > >>>>>>> nodes > >>>>>>>>> but > >>>>>>>>>> when I tried running I got this error: > >>>>>>>>>> > >>>>>>>>>> Exception in thread "main" java.io.IOException: No FileSystem > >>> for > >>>>>>>>> scheme: > >>>>>>>>>> hdfs > >>>>>>>>>> at > >>>>>>>>> > >>>>> > >>>> > >>>>> > org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2421) > >>>>>>>>>> at > >>>>>>>>> > >>>>> > >>>>> > org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2428) > >>>>>>>>>> at > >>> org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88) > >>>>>>>>>> at > >>>>>>>>> > >>>>> > >>>>> > org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467) > >>>>>>>>>> at > >>>> org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449) > >>>>>>>>>> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367) > >>>>>>>>>> at org.apache.hadoop.fs.Path.getFileSystem(Path.java:287) > >>>>>>>>>> at > >>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>> > org.apache.samza.job.yarn.ClientHelper.submitApplication(ClientHelper.s > >>>>>>>>> ca > >>>>>>>>> la:111) > >>>>>>>>>> at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:55) > >>>>>>>>>> at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:48) > >>>>>>>>>> at org.apache.samza.job.JobRunner.run(JobRunner.scala:62) > >>>>>>>>>> at org.apache.samza.job.JobRunner$.main(JobRunner.scala:37) > >>>>>>>>>> at org.apache.samza.job.JobRunner.main(JobRunner.scala) > >>>>>>>>>> > >>>>>>>>>> > >>>>>>>>>> This is my yarn.package.path config: > >>>>>>>>>> > >>>>>>>>>> > >>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>> > yarn.package.path=hdfs://telles-master-samza:50070/samza-job-package-0 > >>>>>>>>> .7 > >>>>>>>>> .0-dist.tar.gz > >>>>>>>>>> > >>>>>>>>>> Thanks in advance > >>>>>>>>>> > >>>>>>>>>> > >>>>>>>>>> > >>>>>>>>>> > >>>>>>>>>> > >>>>>>>>>> On Mon, Aug 11, 2014 at 3:00 PM, Yan Fang < > >>> [email protected]> > >>>>>>>>> wrote: > >>>>>>>>>> > >>>>>>>>>>> Hi Telles, > >>>>>>>>>>> > >>>>>>>>>>> In terms of "*I tried pushing the tar file to HDFS but I got > >>> an > >>>>>>>>> error > >>>>>>>>>> from > >>>>>>>>>>> hadoop saying that it couldn’t find core-site.xml file*.", I > >>>>>>> guess > >>>>>>>>> you > >>>>>>>>>> set > >>>>>>>>>>> the HADOOP_CONF_DIR variable and made it point to > >>>> ~/.samza/conf. > >>>>>>> You > >>>>>>>>> can > >>>>>>>>>> do > >>>>>>>>>>> 1) make the HADOOP_CONF_DIR point to the directory where your > >>>>>>> conf > >>>>>>>>> files > >>>>>>>>>>> are, such as /etc/hadoop/conf. Or 2) copy the config files to > >>>>>>>>>>> ~/.samza/conf. Thank you, > >>>>>>>>>>> > >>>>>>>>>>> Cheer, > >>>>>>>>>>> > >>>>>>>>>>> Fang, Yan > >>>>>>>>>>> [email protected] > >>>>>>>>>>> +1 (206) 849-4108 > >>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>>> On Mon, Aug 11, 2014 at 7:40 AM, Chris Riccomini < > >>>>>>>>>>> [email protected]> wrote: > >>>>>>>>>>> > >>>>>>>>>>>> Hey Telles, > >>>>>>>>>>>> > >>>>>>>>>>>> To get YARN working with the HTTP file system, you need to > >>>>>>> follow > >>>>>>>>> the > >>>>>>>>>>>> instructions on: > >>>>>>>>>>>> > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>>>> > >>>>> > >>>> > >>> > http://samza.incubator.apache.org/learn/tutorials/0.7.0/run-in-multi-node > >>>>>>>>> -y > >>>>>>>>>>>> arn.html > >>>>>>>>>>>> > >>>>>>>>>>>> > >>>>>>>>>>>> In the "Set Up Http Filesystem for YARN" section. > >>>>>>>>>>>> > >>>>>>>>>>>> You shouldn't need to compile anything (no Gradle, which is > >>>>>>> what > >>>>>>>>> your > >>>>>>>>>>>> stack trace is showing). This setup should be done for all > >>> of > >>>>>>> the > >>>>>>>>> NMs, > >>>>>>>>>>>> since they will be the ones downloading your job's package > >>>>>>> (from > >>>>>>>>>>>> yarn.package.path). > >>>>>>>>>>>> > >>>>>>>>>>>> Cheers, > >>>>>>>>>>>> Chris > >>>>>>>>>>>> > >>>>>>>>>>>> On 8/9/14 9:44 PM, "Telles Nobrega" < > >>> [email protected] > >>>>> > >>>>>>>>> wrote: > >>>>>>>>>>>> > >>>>>>>>>>>>> Hi again, I tried installing the scala libs but the Http > >>>>>>> problem > >>>>>>>>> still > >>>>>>>>>>>>> occurs. I realised that I need to compile incubator samza > >>> in > >>>>>>> the > >>>>>>>>>>> machines > >>>>>>>>>>>>> that I¹m going to run the jobs, but the compilation fails > >>>> with > >>>>>>>>> this > >>>>>>>>>> huge > >>>>>>>>>>>>> message: > >>>>>>>>>>>>> > >>>>>>>>>>>>> # > >>>>>>>>>>>>> # There is insufficient memory for the Java Runtime > >>>>>>> Environment > >>>>>>>>> to > >>>>>>>>>>>>> continue. > >>>>>>>>>>>>> # Native memory allocation (malloc) failed to allocate > >>>>>>> 3946053632 > >>>>>>>>>> bytes > >>>>>>>>>>>>> for committing reserved memory. > >>>>>>>>>>>>> # An error report file with more information is saved as: > >>>>>>>>>>>>> # > >>>> /home/ubuntu/incubator-samza/samza-kafka/hs_err_pid2506.log > >>>>>>>>>>>>> Could not write standard input into: Gradle Worker 13. > >>>>>>>>>>>>> java.io.IOException: Broken pipe > >>>>>>>>>>>>> at java.io.FileOutputStream.writeBytes(Native > >>> Method) > >>>>>>>>>>>>> at > >>>>>>>>> java.io.FileOutputStream.write(FileOutputStream.java:345) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>> > >>>>> > >>>>> > java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>> java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.process.internal.streams.ExecOutputHandleRunner.run(ExecOut > >>>>>>>>>> pu > >>>>>>>>>> tH > >>>>>>>>>>>>> andleRunner.java:53) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecuto > >>>>>>>>>> rI > >>>>>>>>>> mp > >>>>>>>>>>>>> l$1.run(DefaultExecutorFactory.java:66) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j > >>>>>>>>>> av > >>>>>>>>>> a: > >>>>>>>>>>>>> 1145) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor. > >>>>>>>>>> ja > >>>>>>>>>> va > >>>>>>>>>>>>> :615) > >>>>>>>>>>>>> at java.lang.Thread.run(Thread.java:744) > >>>>>>>>>>>>> Process 'Gradle Worker 13' finished with non-zero exit > >>>> value 1 > >>>>>>>>>>>>> org.gradle.process.internal.ExecException: Process 'Gradle > >>>>>>> Worker > >>>>>>>>> 13' > >>>>>>>>>>>>> finished with non-zero exit value 1 > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.process.internal.DefaultExecHandle$ExecResultImpl.assertNor > >>>>>>>>>> ma > >>>>>>>>>> lE > >>>>>>>>>>>>> xitValue(DefaultExecHandle.java:362) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.process.internal.DefaultWorkerProcess.onProcessStop(Default > >>>>>>>>>> Wo > >>>>>>>>>> rk > >>>>>>>>>>>>> erProcess.java:89) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.process.internal.DefaultWorkerProcess.access$000(DefaultWor > >>>>>>>>>> ke > >>>>>>>>>> rP > >>>>>>>>>>>>> rocess.java:33) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.process.internal.DefaultWorkerProcess$1.executionFinished(D > >>>>>>>>>> ef > >>>>>>>>>> au > >>>>>>>>>>>>> ltWorkerProcess.java:55) > >>>>>>>>>>>>> at > >>>> sun.reflect.NativeMethodAccessorImpl.invoke0(Native > >>>>>>>>> Method) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j > >>>>>>>>>> av > >>>>>>>>>> a: > >>>>>>>>>>>>> 57) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccess > >>>>>>>>>> or > >>>>>>>>>> Im > >>>>>>>>>>>>> pl.java:43) > >>>>>>>>>>>>> at java.lang.reflect.Method.invoke(Method.java:606) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDi > >>>>>>>>>> sp > >>>>>>>>>> at > >>>>>>>>>>>>> ch.java:35) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDi > >>>>>>>>>> sp > >>>>>>>>>> at > >>>>>>>>>>>>> ch.java:24) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java: > >>>>>>>>>> 81 > >>>>>>>>>> ) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java: > >>>>>>>>>> 30 > >>>>>>>>>> ) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.messaging.dispatch.ProxyDispatchAdapter$DispatchingInvocati > >>>>>>>>>> on > >>>>>>>>>> Ha > >>>>>>>>>>>>> ndler.invoke(ProxyDispatchAdapter.java:93) > >>>>>>>>>>>>> at com.sun.proxy.$Proxy46.executionFinished(Unknown > >>>>>>>>> Source) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.process.internal.DefaultExecHandle.setEndStateInfo(DefaultE > >>>>>>>>>> xe > >>>>>>>>>> cH > >>>>>>>>>>>>> andle.java:212) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.process.internal.DefaultExecHandle.finished(DefaultExecHand > >>>>>>>>>> le > >>>>>>>>>> .j > >>>>>>>>>>>>> ava:309) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.process.internal.ExecHandleRunner.completed(ExecHandleRunne > >>>>>>>>>> r. > >>>>>>>>>> ja > >>>>>>>>>>>>> va:108) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.process.internal.ExecHandleRunner.run(ExecHandleRunner.java > >>>>>>>>>> :8 > >>>>>>>>>> 8) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecuto > >>>>>>>>>> rI > >>>>>>>>>> mp > >>>>>>>>>>>>> l$1.run(DefaultExecutorFactory.java:66) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j > >>>>>>>>>> av > >>>>>>>>>> a: > >>>>>>>>>>>>> 1145) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor. > >>>>>>>>>> ja > >>>>>>>>>> va > >>>>>>>>>>>>> :615) > >>>>>>>>>>>>> at java.lang.Thread.run(Thread.java:744) > >>>>>>>>>>>>> OpenJDK 64-Bit Server VM warning: INFO: > >>>>>>>>>>>>> os::commit_memory(0x000000070a6c0000, 3946053632, 0) > >>> failed; > >>>>>>>>>>>>> error='Cannot allocate memory' (errno=12) > >>>>>>>>>>>>> # > >>>>>>>>>>>>> # There is insufficient memory for the Java Runtime > >>>>>>> Environment > >>>>>>>>> to > >>>>>>>>>>>>> continue. > >>>>>>>>>>>>> # Native memory allocation (malloc) failed to allocate > >>>>>>> 3946053632 > >>>>>>>>>> bytes > >>>>>>>>>>>>> for committing reserved memory. > >>>>>>>>>>>>> # An error report file with more information is saved as: > >>>>>>>>>>>>> # > >>>> /home/ubuntu/incubator-samza/samza-kafka/hs_err_pid2518.log > >>>>>>>>>>>>> Could not write standard input into: Gradle Worker 14. > >>>>>>>>>>>>> java.io.IOException: Broken pipe > >>>>>>>>>>>>> at java.io.FileOutputStream.writeBytes(Native > >>> Method) > >>>>>>>>>>>>> at > >>>>>>>>> java.io.FileOutputStream.write(FileOutputStream.java:345) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>> > >>>>> > >>>>> > java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:82) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>> java.io.BufferedOutputStream.flush(BufferedOutputStream.java:140) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.process.internal.streams.ExecOutputHandleRunner.run(ExecOut > >>>>>>>>>> pu > >>>>>>>>>> tH > >>>>>>>>>>>>> andleRunner.java:53) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecuto > >>>>>>>>>> rI > >>>>>>>>>> mp > >>>>>>>>>>>>> l$1.run(DefaultExecutorFactory.java:66) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j > >>>>>>>>>> av > >>>>>>>>>> a: > >>>>>>>>>>>>> 1145) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor. > >>>>>>>>>> ja > >>>>>>>>>> va > >>>>>>>>>>>>> :615) > >>>>>>>>>>>>> at java.lang.Thread.run(Thread.java:744) > >>>>>>>>>>>>> Process 'Gradle Worker 14' finished with non-zero exit > >>>> value 1 > >>>>>>>>>>>>> org.gradle.process.internal.ExecException: Process 'Gradle > >>>>>>> Worker > >>>>>>>>> 14' > >>>>>>>>>>>>> finished with non-zero exit value 1 > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.process.internal.DefaultExecHandle$ExecResultImpl.assertNor > >>>>>>>>>> ma > >>>>>>>>>> lE > >>>>>>>>>>>>> xitValue(DefaultExecHandle.java:362) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.process.internal.DefaultWorkerProcess.onProcessStop(Default > >>>>>>>>>> Wo > >>>>>>>>>> rk > >>>>>>>>>>>>> erProcess.java:89) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.process.internal.DefaultWorkerProcess.access$000(DefaultWor > >>>>>>>>>> ke > >>>>>>>>>> rP > >>>>>>>>>>>>> rocess.java:33) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.process.internal.DefaultWorkerProcess$1.executionFinished(D > >>>>>>>>>> ef > >>>>>>>>>> au > >>>>>>>>>>>>> ltWorkerProcess.java:55) > >>>>>>>>>>>>> at > >>>> sun.reflect.NativeMethodAccessorImpl.invoke0(Native > >>>>>>>>> Method) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.j > >>>>>>>>>> av > >>>>>>>>>> a: > >>>>>>>>>>>>> 57) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccess > >>>>>>>>>> or > >>>>>>>>>> Im > >>>>>>>>>>>>> pl.java:43) > >>>>>>>>>>>>> at java.lang.reflect.Method.invoke(Method.java:606) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDi > >>>>>>>>>> sp > >>>>>>>>>> at > >>>>>>>>>>>>> ch.java:35) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDi > >>>>>>>>>> sp > >>>>>>>>>> at > >>>>>>>>>>>>> ch.java:24) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java: > >>>>>>>>>> 81 > >>>>>>>>>> ) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.listener.BroadcastDispatch.dispatch(BroadcastDispatch.java: > >>>>>>>>>> 30 > >>>>>>>>>> ) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.messaging.dispatch.ProxyDispatchAdapter$DispatchingInvocati > >>>>>>>>>> on > >>>>>>>>>> Ha > >>>>>>>>>>>>> ndler.invoke(ProxyDispatchAdapter.java:93) > >>>>>>>>>>>>> at com.sun.proxy.$Proxy46.executionFinished(Unknown > >>>>>>>>> Source) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.process.internal.DefaultExecHandle.setEndStateInfo(DefaultE > >>>>>>>>>> xe > >>>>>>>>>> cH > >>>>>>>>>>>>> andle.java:212) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.process.internal.DefaultExecHandle.finished(DefaultExecHand > >>>>>>>>>> le > >>>>>>>>>> .j > >>>>>>>>>>>>> ava:309) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.process.internal.ExecHandleRunner.completed(ExecHandleRunne > >>>>>>>>>> r. > >>>>>>>>>> ja > >>>>>>>>>>>>> va:108) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.process.internal.ExecHandleRunner.run(ExecHandleRunner.java > >>>>>>>>>> :8 > >>>>>>>>>> 8) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecuto > >>>>>>>>>> rI > >>>>>>>>>> mp > >>>>>>>>>>>>> l$1.run(DefaultExecutorFactory.java:66) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.j > >>>>>>>>>> av > >>>>>>>>>> a: > >>>>>>>>>>>>> 1145) > >>>>>>>>>>>>> at > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>> > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor. > >>>>>>>>>> ja > >>>>>>>>>> va > >>>>>>>>>>>>> :615) > >>>>>>>>>>>>> at java.lang.Thread.r > >>>>>>>>>>>>> > >>>>>>>>>>>>> Do I need more memory for my machines? Each already has > >>>> 4GB. I > >>>>>>>>> really > >>>>>>>>>>>>> need to have this running. I¹m not sure which way is best > >>>>>>> http or > >>>>>>>>> hdfs > >>>>>>>>>>>>> which one you suggest and how can i solve my problem for > >>>> each > >>>>>>>>> case. > >>>>>>>>>>>>> > >>>>>>>>>>>>> Thanks in advance and sorry for bothering this much. > >>>>>>>>>>>>> On 10 Aug 2014, at 00:20, Telles Nobrega > >>>>>>>>> <[email protected]> > >>>>>>>>>>> wrote: > >>>>>>>>>>>>> > >>>>>>>>>>>>>> Hi Chris, now I have the tar file in my RM machine, and > >>>> the > >>>>>>>>> yarn > >>>>>>>>>> path > >>>>>>>>>>>>>> points to it. I changed the core-site.xml to use > >>>>>>> HttpFileSystem > >>>>>>>>>> instead > >>>>>>>>>>>>>> of HDFS now it is failing with > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> Application application_1407640485281_0001 failed 2 > >>> times > >>>>>>> due > >>>>>>>>> to > >>>>>>>>> AM > >>>>>>>>>>>>>> Container for appattempt_1407640485281_0001_000002 exited > >>>>>>> with > >>>>>>>>>>>>>> exitCode:-1000 due to: java.lang.ClassNotFoundException: > >>>>>>> Class > >>>>>>>>>>>>>> org.apache.samza.util.hadoop.HttpFileSystem not found > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> I think I can solve this just installing scala files > >>> from > >>>>>>> the > >>>>>>>>> samza > >>>>>>>>>>>>>> tutorial, can you confirm that? > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> On 09 Aug 2014, at 08:34, Telles Nobrega > >>>>>>>>> <[email protected] > >>>>>>>>>> > >>>>>>>>>>>>>> wrote: > >>>>>>>>>>>>>> > >>>>>>>>>>>>>>> Hi Chris, > >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> I think the problem is that I forgot to update the > >>>>>>>>>> yarn.job.package. > >>>>>>>>>>>>>>> I will try again to see if it works now. > >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> I have one more question, how can I stop (command line) > >>>> the > >>>>>>>>> jobs > >>>>>>>>>>>>>>> running in my topology, for the experiment that I will > >>>> run, > >>>>>>> I > >>>>>>>>> need > >>>>>>>>>> to > >>>>>>>>>>>>>>> run the same job in 4 minutes intervals. So I need to > >>> kill > >>>>>>> it, > >>>>>>>>> clean > >>>>>>>>>>>>>>> the kafka topics and rerun. > >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> Thanks in advance. > >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> On 08 Aug 2014, at 12:41, Chris Riccomini > >>>>>>>>>>>>>>> <[email protected]> wrote: > >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> Hey Telles, > >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> Do I need to have the job folder on each machine in > >>> my > >>>>>>>>> cluster? > >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> No, you should not need to do this. There are two ways > >>>> to > >>>>>>>>> deploy > >>>>>>>>>>> your > >>>>>>>>>>>>>>>> tarball to the YARN grid. One is to put it in HDFS, > >>> and > >>>>>>> the > >>>>>>>>> other > >>>>>>>>>> is > >>>>>>>>>>>>>>>> to > >>>>>>>>>>>>>>>> put it on an HTTP server. The link to running a Samza > >>>> job > >>>>>>> in > >>>>>>>>> a > >>>>>>>>>>>>>>>> multi-node > >>>>>>>>>>>>>>>> YARN cluster describes how to do both (either HTTP > >>>> server > >>>>>>> or > >>>>>>>>>> HDFS). > >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> In both cases, once the tarball is put in on the > >>>> HTTP/HDFS > >>>>>>>>>>> server(s), > >>>>>>>>>>>>>>>> you > >>>>>>>>>>>>>>>> must update yarn.package.path to point to it. From > >>>> there, > >>>>>>> the > >>>>>>>>> YARN > >>>>>>>>>>> NM > >>>>>>>>>>>>>>>> should download it for you automatically when you > >>> start > >>>>>>> your > >>>>>>>>> job. > >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> * Can you send along a paste of your job config? > >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> Cheers, > >>>>>>>>>>>>>>>> Chris > >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> On 8/8/14 8:04 AM, "Claudio Martins" > >>>>>>>>> <[email protected]> > >>>>>>>>>>> wrote: > >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> Hi Telles, it looks to me that you forgot to update > >>> the > >>>>>>>>>>>>>>>>> "yarn.package.path" > >>>>>>>>>>>>>>>>> attribute in your config file for the task. > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> - Claudio Martins > >>>>>>>>>>>>>>>>> Head of Engineering > >>>>>>>>>>>>>>>>> MobileAware USA Inc. / www.mobileaware.com > >>>>>>>>>>>>>>>>> office: +1 617 986 5060 / mobile: +1 617 480 5288 > >>>>>>>>>>>>>>>>> linkedin: www.linkedin.com/in/martinsclaudio > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> On Fri, Aug 8, 2014 at 10:55 AM, Telles Nobrega > >>>>>>>>>>>>>>>>> <[email protected]> > >>>>>>>>>>>>>>>>> wrote: > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> Hi, > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> this is my first time trying to run a job on a > >>>> multinode > >>>>>>>>>>>>>>>>>> environment. I > >>>>>>>>>>>>>>>>>> have the cluster set up, I can see in the GUI that > >>> all > >>>>>>>>> nodes > >>>>>>>>> are > >>>>>>>>>>>>>>>>>> working. > >>>>>>>>>>>>>>>>>> Do I need to have the job folder on each machine in > >>> my > >>>>>>>>> cluster? > >>>>>>>>>>>>>>>>>> - The first time I tried running with the job on the > >>>>>>>>> namenode > >>>>>>>>>>>>>>>>>> machine > >>>>>>>>>>>>>>>>>> and > >>>>>>>>>>>>>>>>>> it failed saying: > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> Application application_1407509228798_0001 failed 2 > >>>>>>> times > >>>>>>>>> due > >>>>>>>>> to > >>>>>>>>>>> AM > >>>>>>>>>>>>>>>>>> Container for appattempt_1407509228798_0001_000002 > >>>>>>> exited > >>>>>>>>> with > >>>>>>>>>>>>>>>>>> exitCode: > >>>>>>>>>>>>>>>>>> -1000 due to: File > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>>>>>>> > file:/home/ubuntu/alarm-samza/samza-job-package/target/samza-job- > >>>>>>>>>>>>>>> pa > >>>>>>>>>>>>>>> ck > >>>>>>>>>>>>>>>>>> age- > >>>>>>>>>>>>>>>>>> 0.7.0-dist.tar.gz > >>>>>>>>>>>>>>>>>> does not exist > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> So I copied the folder to each machine in my cluster > >>>> and > >>>>>>>>> got > >>>>>>>>>> this > >>>>>>>>>>>>>>>>>> error: > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> Application application_1407509228798_0002 failed 2 > >>>>>>> times > >>>>>>>>> due > >>>>>>>>> to > >>>>>>>>>>> AM > >>>>>>>>>>>>>>>>>> Container for appattempt_1407509228798_0002_000002 > >>>>>>> exited > >>>>>>>>> with > >>>>>>>>>>>>>>>>>> exitCode: > >>>>>>>>>>>>>>>>>> -1000 due to: Resource > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>> > >>>>>>> > >>>>> > >>>> > >>>>>>>>>>>>> > file:/home/ubuntu/alarm-samza/samza-job-package/target/samza-job- > >>>>>>>>>>>>>>> pa > >>>>>>>>>>>>>>> ck > >>>>>>>>>>>>>>>>>> age- > >>>>>>>>>>>>>>>>>> 0.7.0-dist.tar.gz > >>>>>>>>>>>>>>>>>> changed on src filesystem (expected 1407509168000, > >>> was > >>>>>>>>>>> 1407509434000 > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> What am I missing? > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> p.s.: I followed this > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> < > >>>>>>>>>>>> > >>>>>>>>> > >>>> https://github.com/yahoo/samoa/wiki/Executing-SAMOA-with-Apache-Samz > >>>>>>>>>>>>>>>>>> a> > >>>>>>>>>>>>>>>>>> tutorial > >>>>>>>>>>>>>>>>>> and this > >>>>>>>>>>>>>>>>>> < > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>> > >>>>>>>>> > >>>>> > http://samza.incubator.apache.org/learn/tutorials/0.7.0/run-in-multi- > >>>>>>>>>>>>>>>>>> node > >>>>>>>>>>>>>>>>>> -yarn.html > >>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> to > >>>>>>>>>>>>>>>>>> set up the cluster. > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> Help is much appreciated. > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> Thanks in advance. > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> -- > >>>>>>>>>>>>>>>>>> ------------------------------------------ > >>>>>>>>>>>>>>>>>> Telles Mota Vidal Nobrega > >>>>>>>>>>>>>>>>>> M.sc. Candidate at UFCG > >>>>>>>>>>>>>>>>>> B.sc. in Computer Science at UFCG > >>>>>>>>>>>>>>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> > >>>>>>>>>>>>>> > >>>>>>>>>>>>> > >>>>>>>>>>>> > >>>>>>>>>>>> > >>>>>>>>>>> > >>>>>>>>>> > >>>>>>>>>> > >>>>>>>>>> > >>>>>>>>>> -- > >>>>>>>>>> ------------------------------------------ > >>>>>>>>>> Telles Mota Vidal Nobrega > >>>>>>>>>> M.sc. Candidate at UFCG > >>>>>>>>>> B.sc. in Computer Science at UFCG > >>>>>>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG > >>>>>>>>>> > >>>>>>>>> > >>>>>>>> > >>>>>>>> > >>>>>>>> > >>>>>>>> -- > >>>>>>>> ------------------------------------------ > >>>>>>>> Telles Mota Vidal Nobrega > >>>>>>>> M.sc. Candidate at UFCG > >>>>>>>> B.sc. in Computer Science at UFCG > >>>>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG > >>>>>>> > >>>>>>> > >>>>>> > >>>>>> > >>>>>> -- > >>>>>> ------------------------------------------ > >>>>>> Telles Mota Vidal Nobrega > >>>>>> M.sc. Candidate at UFCG > >>>>>> B.sc. in Computer Science at UFCG > >>>>>> Software Engineer at OpenStack Project - HP/LSD-UFCG > >>>>> > >>>>> > >>>> > >>> > >>> > >>> > >>> -- > >>> ------------------------------------------ > >>> Telles Mota Vidal Nobrega > >>> M.sc. Candidate at UFCG > >>> B.sc. in Computer Science at UFCG > >>> Software Engineer at OpenStack Project - HP/LSD-UFCG > >>> > > > >
