Hey Sonali, You should not have to add yarn.rm.hostname to your configuration. You have applied the wrong patch. Please apply this patch:
https://issues.apache.org/jira/secure/attachment/12634493/SAMZA-182.1.patch Do not apply https://issues.apache.org/jira/secure/attachment/12634200/yarn_remote_rm.pa tch, which is the one that adds yarn.rm.hostname. Regarding the failed containers, can you take a look at the NM's logs, and see what it says there? Cheers, Chris On 3/13/14 12:00 PM, "[email protected]" <[email protected]> wrote: >Hi Chris. > >I installed the patch and was able to submit the job. (I had to add >yarn.rm.hostname in the wikipedia-feed.properties file). The job started >running and then I got this error: >SamzaAppMasterTaskManager [INFO] Got an exit code of -100. This means >that container container_1391637982288_0040_01_000043 was killed by YARN, >either due to being released by the application master or being 'lost' >due to node failures etc. >2014-03-13 11:53:03 SamzaAppMasterTaskManager [INFO] Got an exit code of >-100. This means that container container_1391637982288_0040_01_000044 >was killed by YARN, either due to being released by the application >master or being 'lost' due to node failures etc. >2014-03-13 11:53:03 SamzaAppMasterTaskManager [INFO] Got an exit code of >-100. This means that container container_1391637982288_0040_01_000045 >was killed by YARN, either due to being released by the application >master or being 'lost' due to node failures etc. >2014-03-13 11:53:03 SamzaAppMasterTaskManager [INFO] Got a container from >YARN ResourceManager: Container: [ContainerId: >container_1391637982288_0040_01_000046, NodeId: svdpdac015:37919, >NodeHttpAddress: svdpdac015:8042, Resource: <memory:1024, vCores:1>, >Priority: 0, Token: Token { kind: ContainerToken, service: >10.1.174.84:37919 }, ] >2014-03-13 11:53:03 SamzaAppMasterTaskManager [INFO] Got an extra >container from YARN ResourceManager: Container: [ContainerId: >container_1391637982288_0040_01_000046, NodeId: svdpdac015:37919, >NodeHttpAddress: svdpdac015:8042, Resource: <memory:1024, vCores:1>, >Priority: 0, Token: Token { kind: ContainerToken, service: >10.1.174.84:37919 }, ] >2014-03-13 11:53:04 SamzaAppMasterTaskManager [INFO] Got an exit code of >-100. This means that container container_1391637982288_0040_01_000046 >was killed by YARN, either due to being released by the application >master or being 'lost' due to node failures etc. >2014-03-13 11:53:24 SamzaAppMasterTaskManager [INFO] Container >container_1391637982288_0040_01_000038 failed with exit code 1 - >Exception from container-launch: >org.apache.hadoop.util.Shell$ExitCodeException: > at org.apache.hadoop.util.Shell.runCommand(Shell.java:464) > at org.apache.hadoop.util.Shell.run(Shell.java:379) > at >org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:589) > at >org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.launchC >ontainer(DefaultContainerExecutor.java:195) > at >org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.Contai >nerLaunch.call(ContainerLaunch.java:283) > at >org.apache.hadoop.yarn.server.nodemanager.containermanager.launcher.Contai >nerLaunch.call(ContainerLaunch.java:79) > at java.util.concurrent.FutureTask.run(FutureTask.java:262) > at >java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java: >1145) > at >java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java >:615) > at java.lang.Thread.run(Thread.java:744) > > >Now, I'm running Hadoop-2.2.0 version of yarn. Do I need to change to >YARN 2.3? > >-----Original Message----- >From: Chris Riccomini [mailto:[email protected]] >Sent: Thursday, March 13, 2014 11:12 AM >To: [email protected] >Subject: Re: Failed to package using mvn > >Hey Sonali, > >As a follow on, here is EXACTLY what I did: > ># start by downloading YARN and setting it up 1. Download YARN 2.3 from >http://mirror.symnds.com/software/Apache/hadoop/common/hadoop-2.3.0/hadoop >- >2.3.0.tar.gz to /tmp >2. cd /tmp >3. tar -xvf hadoop-2.3.0.tar.gz >4. cd hadoop-2.3.0 >5. export HADOOP_YARN_HOME=$(pwd) >6. mkdir conf >7. export HADOOP_CONF_DIR=$HADOOP_YARN_HOME/conf >8. cp ./etc/hadoop/yarn-site.xml conf >9. vi conf/yarn-site.xml >10. Add this property to yarn-site.xml: > > <property> > <name>yarn.resourcemanager.hostname</name> > <!-- hostname that is accessible from all NMs --> > <value>criccomi-mn</value> > </property> > > >11. curl >http://svn.apache.org/viewvc/hadoop/common/trunk/hadoop-yarn-project/hadoo >p >-yarn/hadoop-yarn-server/hadoop-yarn-server-tests/src/test/resources/capac >i >ty-scheduler.xml?view=co > conf/capacity-scheduler.xml > ># setup http filesystem for YARN (you can skip this and follow SAMZA-181 >instead, if you are using HDFS) 12. cd /tmp 13. curl >http://www.scala-lang.org/files/archive/scala-2.10.3.tgz > >scala-2.10.3.tgz 14. tar -xvf scala-2.10.3.tgz 15. cp >/tmp/scala-2.10.3/lib/scala-compiler.jar >$HADOOP_YARN_HOME/share/hadoop/hdfs/lib >16. cp /tmp/scala-2.10.3/lib/scala-library.jar >$HADOOP_YARN_HOME/share/hadoop/hdfs/lib >17. curl >http://search.maven.org/remotecontent?filepath=org/clapper/grizzled-slf4j_ >2 >.10/1.0.1/grizzled-slf4j_2.10-1.0.1.jar > >$HADOOP_YARN_HOME/share/hadoop/hdfs/lib/grizzled-slf4j_2.10-1.0.1.jar >18. vi $HADOOP_YARN_HOME/conf/core-site.xml > ><?xml-stylesheet type="text/xsl" href="configuration.xsl"?> ><configuration> <property> <name>fs.http.impl</name> ><value>org.apache.samza.util.hadoop.HttpFileSystem</value> ></property> ></configuration> > > >19. Copy the Hadoop directory to all slave nodes (172.21.100.35, in my >case): > > scp -r . 172.21.100.35:/tmp/hadoop-2.3.0 > >20. echo 172.21.100.35 > conf/slaves >21. sbin/start-yarn.sh >22. If you get "172.21.100.35: Error: JAVA_HOME is not set and could not >be found.", you'll need to add a conf/hadoop-env.sh file to the machine >with the failure (172.21.100.35, in this case), which has "export >JAVA_HOME=/export/apps/jdk/JDK-1_6_0_27" (or wherever your JAVA_HOME >actually is). >23. Validate that your nodes are up by visiting >http://criccomi-mn:8088/cluster/nodes > > ># now we more or less follow the hello-samza steps. >24. cd /tmp >25. git clone http://git-wip-us.apache.org/repos/asf/incubator-samza.git >26. cd incubator-samza >27. curl >https://issues.apache.org/jira/secure/attachment/12634493/SAMZA-182.1.patc >h > > SAMZA-182.1.patch >28. git apply SAMZA-182.1.patch >29. ./gradlew clean publishToMavenLocal >30. cd .. >31. git clone git://github.com/linkedin/hello-samza.git >32. cd hello-samza >33. vi samza-job-package/src/main/config/wikipedia-feed.properties >34. Change the yarn.package.path property to be: > > >yarn.package.path=http://criccomi-mn:8000/samza-job-package/target/samza-j >o >b-package-0.7.0-dist.tar.gz > > >35. mvn clean package >36. mkdir -p deploy/samza >37. tar -xvf >./samza-job-package/target/samza-job-package-0.7.0-dist.tar.gz -C >deploy/samza 38. Open a new terminal, and cd /tmp/hello-samza && python >-m SimpleHTTPServer 39. Go back to the original terminal (not the one >running the HTTP server) 40. deploy/samza/bin/run-job.sh >--config-factory=org.apache.samza.config.factories.PropertiesConfigFactory >--config-path=file://$PWD/deploy/samza/config/wikipedia-feed.properties >41. Go to http://criccomi-mn:8088 and find the wikipedia-feed job. Click >on the ApplicationMaster link to see that it's running. > > >I plan to write a tutorial that formalizes this. > >Cheers, >Chris > >On 3/13/14 11:03 AM, "Chris Riccomini" <[email protected]> wrote: > >>Hey Sonali, >> >>Please have a look at: >> >> https://issues.apache.org/jira/browse/SAMZA-182 >> >>I have posted a fix there. I have successfully downloaded YARN, setup a >>two node grid from scratch, and run hello-samza on it with the patch on >>SAMZA-182. Can you give that a shot for me? >> >>Thanks for your patience! >> >>Cheers, >>Chris >> >>On 3/13/14 10:58 AM, "[email protected]" >><[email protected]> wrote: >> >>>Hi Chris, >>> >>>I checked my .bashrc. The variable was set on one of the NMs and not on >>>the other. I made the change and restarted the scripts. I still get the >>>same error. >>> >>>Also in my stderr I get: >>>Null identity service, trying login service: null >>>Finding identity service: null >>>Null identity service, trying login service: null >>>Finding identity service: null >>> >>>-----Original Message----- >>>From: Chris Riccomini [mailto:[email protected]] >>>Sent: Wednesday, March 12, 2014 7:59 PM >>>To: [email protected] >>>Subject: Re: Failed to package using mvn >>> >>>Hey Guys, >>> >>>I was able to reproduce this problem. >>> >>>I was also able to fix it (without the patch in SAMZA-182). All I needed >>>to do was update ~/.bashrc on my NM's box to have: >>> >>> export YARN_HOME=/tmp/hadoop-2.3.0 >>> >>>It appears that the YARN environment variables are somehow getting lost >>>or not forwarded from the NM to the AM. Adding this bashrc setting makes >>>sure that the NM gets them. >>> >>> >>>I have a feeling upgrading Samza to YARN 2.3.0 will fix this, but I >>>haven't validated yet. I will continue to investigate tomorrow. >>> >>>Cheers, >>>Chris >>> >>>On 3/12/14 6:43 PM, "Yan Fang" <[email protected]> wrote: >>> >>>>I guess Sonali has the problem is because his NMs do not read the >>>>YARN_HOME variable. That may be because the NM machine does not have >>>>YARN_HOME set when the NM starts. >>>> >>>>Check this https://issues.apache.org/jira/browse/SAMZA-182 >>>> >>>>Thanks, >>>> >>>>Yan Fang >>>> >>>>> On Mar 12, 2014, at 6:14 PM, Chris Riccomini >>>>><[email protected]> >>>>>wrote: >>>>> >>>>> Hey Sonali, >>>>> >>>>> I am unfamiliar with the start-yarn.sh. Looking at: >>>>> >>>>> >>>>> >>>>>https://svn.apache.org/viewvc/hadoop/common/trunk/hadoop-yarn-project/ >>>>>had >>>>>oo >>>>> p-yarn/bin/stop-yarn.sh?revision=1370666&view=markup >>>>> >>>>> What version of YARN are you using? >>>>> >>>>> Cheers, >>>>> Chris >>>>> >>>>> On 3/12/14 5:56 PM, "[email protected]" >>>>> <[email protected]> wrote: >>>>> >>>>>> Hey Chris, >>>>>> >>>>>> Yes, I have YARN_HOME set in all the NMs pointing to the right >>>>>>directories. I also made sure the yarn-site.xml file has the hostname >>>>>>set. >>>>>> >>>>>> I start yarn using start.yarn.sh in the RM and that automatically >>>>>>starts the NMs on the slave nodes. Is that the right way to do it? >>>>>> >>>>>> -----Original Message----- >>>>>> From: Chris Riccomini [mailto:[email protected]] >>>>>> Sent: Wednesday, March 12, 2014 5:52 PM >>>>>> To: [email protected] >>>>>> Subject: Re: Failed to package using mvn >>>>>> >>>>>> Hey Sonali, >>>>>> >>>>>> OK, so we've validated that the NMs are able to connect, which means >>>>>>they can see the yarn-site.xml. >>>>>> >>>>>> How are you starting your NMs? Are you running: >>>>>> >>>>>> export YARN_HOME=/path/to/yarn/home >>>>>> >>>>>> In the CLI before starting the NM? >>>>>> >>>>>> For reference, we run: >>>>>> >>>>>> export YARN_HOME=/path/to/our/yarn-home export >>>>>> YARN_CONF_DIR=$YARN_HOME/conf >>>>>> >>>>>> bin/yarn nodemanager >>>>>> >>>>>> With YARN_HOME pointing to a directory that has a subdirectory >>>>>> called "conf" in it, which has a yarn-site.xml in it: >>>>>> >>>>>> /path/to/our/yarn-home/conf/yarn-site.xml >>>>>> >>>>>> This yarn-site.xml has yarn.resourcemanager.hostname set to the IP >>>>>> (or >>>>>> hostname) of the resource manager: >>>>>> >>>>>> <property> >>>>>> <name>yarn.resourcemanager.hostname</name> >>>>>> <value>123.456.789.123</value> >>>>>> </property> >>>>>> >>>>>> >>>>>> Cheers, >>>>>> Chris >>>>>> >>>>>> On 3/12/14 5:33 PM, "[email protected]" >>>>>> <[email protected]> wrote: >>>>>> >>>>>>> I see two active nodes (I have 2 NMs running) >>>>>>> >>>>>>> -----Original Message----- >>>>>>> From: Chris Riccomini [mailto:[email protected]] >>>>>>> Sent: Wednesday, March 12, 2014 5:24 PM >>>>>>> To: [email protected] >>>>>>> Subject: Re: Failed to package using mvn >>>>>>> >>>>>>> Hey Sonali, >>>>>>> >>>>>>> Can you go to your ResourceManager's UI, and tell me how many >>>>>>> active nodes you see? This should be under the "active nodes" >>>>>>>heading. >>>>>>> >>>>>>> It sounds like the SamzaAppMaster is not getting the resource >>>>>>> manager host/port from the yarn-site.xml. Usually this is due to >>>>>>> not exporting YARN_HOME on the NodeManager before starting it. >>>>>>> >>>>>>> Cheers, >>>>>>> Chris >>>>>>> >>>>>>> On 3/12/14 5:21 PM, "[email protected]" >>>>>>> <[email protected]> wrote: >>>>>>> >>>>>>>> Okay so I was able to submit the job: >>>>>>>> >>>>>>>> In the nodemanager I get this error: Specifically it's trying to >>>>>>>> connect to 0.0.0.0/8032 instead of the IP I have specified in the >>>>>>>> yarn-site.xml file >>>>>>>> >>>>>>>> 2014-03-12 17:04:47 SamzaAppMaster$ [INFO] got container id: >>>>>>>> container_1391637982288_0033_01_000001 >>>>>>>> 2014-03-12 17:04:47 SamzaAppMaster$ [INFO] got app attempt id: >>>>>>>> appattempt_1391637982288_0033_000001 >>>>>>>> 2014-03-12 17:04:47 SamzaAppMaster$ [INFO] got node manager host: >>>>>>>> svdpdac001.techlabs.accenture.com >>>>>>>> 2014-03-12 17:04:47 SamzaAppMaster$ [INFO] got node manager port: >>>>>>>> 38218 >>>>>>>> 2014-03-12 17:04:47 SamzaAppMaster$ [INFO] got node manager http >>>>>>>>port: >>>>>>>> 8042 >>>>>>>> 2014-03-12 17:04:47 SamzaAppMaster$ [INFO] got config: >>>>>>>> >>>>>>>>{task.inputs=wikipedia.#en.wikipedia,wikipedia.#en.wiktionary,wikip >>>>>>>>edi a .#e n.wikinews, systems.wikipedia.host=irc.wikimedia.org, >>>>>>>> systems.kafka.producer.batch.num.messages=1, >>>>>>>> job.factory.class=org.apache.samza.job.yarn.YarnJobFactory, >>>>>>>> systems.wikipedia.port=6667, >>>>>>>> systems.kafka.producer.producer.type=sync, >>>>>>>> job.name=wikipedia-feed, >>>>>>>> >>>>>>>>systems.kafka.consumer.zookeeper.connect=svdpdac013.techlabs.accent >>>>>>>>ure >>>>>>>>. >>>>>>>> com :2181/, systems.kafka.samza.msg.serde=json, >>>>>>>> >>>>>>>>serializers.registry.json.class=org.apache.samza.serializers.JsonSe >>>>>>>>rde >>>>>>>> F >>>>>>>> act ory, >>>>>>>> task.class=samza.examples.wikipedia.task.WikipediaFeedStreamTask, >>>>>>>> >>>>>>>>yarn.package.path=hdfs://10.1.174.85:9000/samza-job-package-0.7.0-d >>>>>>>>ist >>>>>>>>. >>>>>>>> tar >>>>>>>> .gz, >>>>>>>> >>>>>>>>systems.wikipedia.samza.factory=samza.examples.wikipedia.system.Wik >>>>>>>>ipe >>>>>>>> d >>>>>>>> iaS >>>>>>>> ystemFactory, >>>>>>>> >>>>>>>>systems.kafka.samza.factory=org.apache.samza.system.kafka.KafkaSyst >>>>>>>>emF >>>>>>>> a >>>>>>>> cto >>>>>>>> ry, >>>>>>>> >>>>>>>>systems.kafka.producer.metadata.broker.list=svdpdac001.techlabs.acc >>>>>>>>ent >>>>>>>> ure >>>>>>>> . >>>>>>>> com:6667,svdpdac015.techlabs.accenture.com:6667} >>>>>>>> 2014-03-12 17:04:48 ClientHelper [INFO] trying to connect to RM >>>>>>>> 0.0.0.0:8032 >>>>>>>> 2014-03-12 17:04:48 NativeCodeLoader [WARN] Unable to load >>>>>>>>native-hadoop library for your platform... using builtin-java >>>>>>>>classes where applicable >>>>>>>> 2014-03-12 17:04:48 RMProxy [INFO] Connecting to ResourceManager >>>>>>>>at >>>>>>>> /0.0.0.0:8032 >>>>>>>> >>>>>>>> -----Original Message----- >>>>>>>> From: Chris Riccomini [mailto:[email protected]] >>>>>>>> Sent: Wednesday, March 12, 2014 4:48 PM >>>>>>>> To: [email protected] >>>>>>>> Subject: Re: Failed to package using mvn >>>>>>>> >>>>>>>> Hey Sonali, >>>>>>>> >>>>>>>> You need to specify a valid HDFS uri. Usually something like: >>>>>>>> >>>>>>>> hdfs://<hdfs name node ip>:<hdfs name node port>/path/to/tgz >>>>>>>> >>>>>>>> Right now, Hadoop is trying to use the package name as the HDFS >>>>>>>>host. >>>>>>>> >>>>>>>> Cheers, >>>>>>>> Chris >>>>>>>> >>>>>>>> On 3/12/14 4:45 PM, "[email protected]" >>>>>>>> <[email protected]> wrote: >>>>>>>> >>>>>>>>> I did and I can now see the hadoop-hdfs jar in /deploy/samza/lib >>>>>>>>> folder. >>>>>>>>> >>>>>>>>> I do get a different error now. >>>>>>>>> >>>>>>>>> I uploaded the samza-job to hdfs and it resides on >>>>>>>>> hdfs://samza-job-package-0.7.0-dist.tar.gz >>>>>>>>> >>>>>>>>> But when I run the job I get this exception: >>>>>>>>> >>>>>>>>> Exception in thread "main" java.lang.IllegalArgumentException: >>>>>>>>> java.net.UnknownHostException: >>>>>>>>>samza-job-package-0.7.0-dist.tar.gz >>>>>>>>> at >>>>>>>>> >>>>>>>>>org.apache.hadoop.security.SecurityUtil.buildTokenService(Security >>>>>>>>>Uti >>>>>>>>>l. >>>>>>>>> jav >>>>>>>>> a:418) >>>>>>>>> at >>>>>>>>> >>>>>>>>>org.apache.hadoop.hdfs.NameNodeProxies.createNonHAProxy(NameNodePr >>>>>>>>>oxi >>>>>>>>> e >>>>>>>>> s >>>>>>>>> .ja >>>>>>>>> va:231) >>>>>>>>> at >>>>>>>>> >>>>>>>>>org.apache.hadoop.hdfs.NameNodeProxies.createProxy(NameNodeProxies >>>>>>>>>.ja >>>>>>>>> v >>>>>>>>> a >>>>>>>>> :13 >>>>>>>>> 9) >>>>>>>>> at >>>>>>>>>org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:510) >>>>>>>>> at >>>>>>>>>org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:453) >>>>>>>>> at >>>>>>>>> >>>>>>>>>org.apache.hadoop.hdfs.DistributedFileSystem.initialize(Distribute >>>>>>>>>dFi >>>>>>>>> l >>>>>>>>> e >>>>>>>>> Sys >>>>>>>>> tem.java:136) >>>>>>>>> at >>>>>>>>> >>>>>>>>>org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2 >>>>>>>>>433 >>>>>>>>>) >>>>>>>>> at >>>>>>>>> org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88) >>>>>>>>> at >>>>>>>>> >>>>>>>>>org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java: >>>>>>>>>246 >>>>>>>>>7) >>>>>>>>> at >>>>>>>>> org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449) >>>>>>>>> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367) >>>>>>>>> at org.apache.hadoop.fs.Path.getFileSystem(Path.java:287) >>>>>>>>> at >>>>>>>>> >>>>>>>>>org.apache.samza.job.yarn.ClientHelper.submitApplication(ClientHel >>>>>>>>>per >>>>>>>>>. >>>>>>>>> s >>>>>>>>> cal >>>>>>>>> a:111) >>>>>>>>> at >>>>>>>>>org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:55) >>>>>>>>> at >>>>>>>>>org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:48) >>>>>>>>> at org.apache.samza.job.JobRunner.run(JobRunner.scala:100) >>>>>>>>> at org.apache.samza.job.JobRunner$.main(JobRunner.scala:75) >>>>>>>>> at org.apache.samza.job.JobRunner.main(JobRunner.scala) >>>>>>>>> Caused by: java.net.UnknownHostException: >>>>>>>>> samza-job-package-0.7.0-dist.tar.gz >>>>>>>>> ... 18 more >>>>>>>>> >>>>>>>>> >>>>>>>>> -----Original Message----- >>>>>>>>> From: Yan Fang [mailto:[email protected]] >>>>>>>>> Sent: Wednesday, March 12, 2014 4:20 PM >>>>>>>>> To: [email protected] >>>>>>>>> Subject: Re: Failed to package using mvn >>>>>>>>> >>>>>>>>> Hi Sonali, >>>>>>>>> >>>>>>>>> One tip you may miss: >>>>>>>>> >>>>>>>>> If you had already run >>>>>>>>> >>>>>>>>> tar -xvf >>>>>>>>> ./samza-job-package/target/samza-job-package-0.7.0-dist.tar.gz >>>>>>>>> -C deploy/samza >>>>>>>>> >>>>>>>>> before you bundled the jar file to tar.gz. Please also remember >>>>>>>>> to put the hdfs jar file to the deploy/samza/lib. >>>>>>>>> >>>>>>>>> Let me know if you miss this step. >>>>>>>>> >>>>>>>>> Thanks, >>>>>>>>> >>>>>>>>> Fang, Yan >>>>>>>>> [email protected] >>>>>>>>> +1 (206) 849-4108 >>>>>>>>> >>>>>>>>> >>>>>>>>> On Wed, Mar 12, 2014 at 4:10 PM, Chris Riccomini >>>>>>>>> <[email protected]>wrote: >>>>>>>>> >>>>>>>>>> Hey Sonali, >>>>>>>>>> >>>>>>>>>> Yan has made a step-by-step tutorial for this. Could you confirm >>>>>>>>>> that you've followed the instructions, and it's still not >>>>>>>>>>working? >>>>>>>>>> >>>>>>>>>> https://issues.apache.org/jira/browse/SAMZA-181 >>>>>>>>>> >>>>>>>>>> Cheers, >>>>>>>>>> Chris >>>>>>>>>> >>>>>>>>>> On 3/12/14 3:12 PM, "[email protected]" >>>>>>>>>> <[email protected]> wrote: >>>>>>>>>> >>>>>>>>>>> So sigh! I had some Kafka issues in-between. That's fixed now. >>>>>>>>>>> >>>>>>>>>>> As suggested, >>>>>>>>>>> >>>>>>>>>>> 1. I made sure the hadoop-hdfs-2.2.0.jar is bundled with the >>>>>>>>>>> samza job tar.gz. >>>>>>>>>>> 2. I added the configuration to implement hdfs in the >>>>>>>>>>> hdfs-site.xml files both on the NMs and in the /conf directory >>>>>>>>>>> for samza >>>>>>>>>>> >>>>>>>>>>> I still get the No Filesystem for scheme :hdfs error. >>>>>>>>>>> >>>>>>>>>>> Is there anything else im missing? >>>>>>>>>>> Thanks, >>>>>>>>>>> Sonali >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -----Original Message----- >>>>>>>>>>> From: Chris Riccomini [mailto:[email protected]] >>>>>>>>>>> Sent: Tuesday, March 11, 2014 8:27 PM >>>>>>>>>>> To: [email protected] >>>>>>>>>>> Subject: Re: Failed to package using mvn >>>>>>>>>>> >>>>>>>>>>> Hey Yan, >>>>>>>>>>> >>>>>>>>>>> This looks great! I added a few requests to the JIRA, if you >>>>>>>>>>> have >>>>>>>>>> time. >>>>>>>>>>> >>>>>>>>>>> Cheers, >>>>>>>>>>> Chris >>>>>>>>>>> >>>>>>>>>>>> On 3/11/14 7:20 PM, "Yan Fang" <[email protected]> wrote: >>>>>>>>>>>> >>>>>>>>>>>> Hi Chris, >>>>>>>>>>>> >>>>>>>>>>>> Has opened an issue >>>>>>>>>>>> SAMZA-181<https://issues.apache.org/jira/browse/SAMZA-181>and >>>>>>>>>>>> also uploaded the patch. Let me know if there is something >>>>>>>>>>>> wrong in my tutorial. Thank you! >>>>>>>>>>>> >>>>>>>>>>>> Cheers, >>>>>>>>>>>> >>>>>>>>>>>> Fang, Yan >>>>>>>>>>>> [email protected] >>>>>>>>>>>> +1 (206) 849-4108 >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Tue, Mar 11, 2014 at 10:40 AM, >>>>>>>>>>>> <[email protected]>wrote: >>>>>>>>>>>> >>>>>>>>>>>>> Thanks Chris, Yan, >>>>>>>>>>>>> >>>>>>>>>>>>> Let me try that. >>>>>>>>>>>>> >>>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>>> From: Chris Riccomini [mailto:[email protected]] >>>>>>>>>>>>> Sent: Tuesday, March 11, 2014 10:22 AM >>>>>>>>>>>>> To: [email protected] >>>>>>>>>>>>> Subject: Re: Failed to package using mvn >>>>>>>>>>>>> >>>>>>>>>>>>> Hey Yan, >>>>>>>>>>>>> >>>>>>>>>>>>> Awesome!The location where you can add your .md is here: >>>>>>>>>>>>> >>>>>>>>>>>>> docs/learn/tutorials/0.7.0/ >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> Here's a link to the code tree: >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> https://git-wip-us.apache.org/repos/asf?p=incubator-samza.git >>>>>>>>>>>>> ;a= >>>>>>>>>>>>> t >>>>>>>>>>>>> r >>>>>>>>>>>>> e >>>>>>>>>>>>> e;f >>>>>>>>>>>>> =do >>>>>>>>>>>>> cs >>>>>>>>>>>>> >>>>>>>>>>>>> /learn/tutorials/0.7.0;h=ef117f4066f14a00f50f0f6fca1790313044 >>>>>>>>>>>>> 831 >>>>>>>>>>>>> 2 >>>>>>>>>>>>> ; >>>>>>>>>>>>> h >>>>>>>>>>>>> b=H >>>>>>>>>>>>> EAD >>>>>>>>>>>>> >>>>>>>>>>>>> You can get the code here: >>>>>>>>>>>>> >>>>>>>>>>>>> git clone >>>>>>>>>>>>> http://git-wip-us.apache.org/repos/asf/incubator-samza.git >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> Once you write the .md, just throw it up on a JIRA, and one >>>>>>>>>>>>> of us can merge it in. >>>>>>>>>>>>> >>>>>>>>>>>>> Re: hdfs-site.xml, ah ha, that's what I figured. This is good >>>>>>>>>>>>> to >>>>>>>>>> know. >>>>>>>>>>>>> So >>>>>>>>>>>>> you just copy your hdfs-site.xml from your NodeManager's conf >>>>>>>>>>>>> directory into your local hdfs-site.xml. >>>>>>>>>>>>> >>>>>>>>>>>>> Cheers, >>>>>>>>>>>>> Chris >>>>>>>>>>>>> >>>>>>>>>>>>>> On 3/11/14 10:16 AM, "Yan Fang" <[email protected]> >>>>>>>>>>>>>>wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>> Hi Chris, >>>>>>>>>>>>>> >>>>>>>>>>>>>> Sure. I just do not know how/where to contribute this >>>>>>>>>>>>>> page...*_* >>>>>>>>>>>>>> >>>>>>>>>>>>>> Oh, I mean the same this as you mentioned in the *Cluster >>>>>>>>>>>>>> Installation*thread: >>>>>>>>>>>>>> >>>>>>>>>>>>>> *"2. Get a copy of one of your NM's yarn-site.xml and put it >>>>>>>>>>>>>> somewhere >>>>>>>>>>>>>> on* >>>>>>>>>>>>>> >>>>>>>>>>>>>> *your desktop (I usually use ~/.yarn/conf/yarn-site.xml). >>>>>>>>>>>>>> Note that there'sa "conf" directory there. This is >>>>>>>>>>>>>> mandatory."* >>>>>>>>>>>>>> >>>>>>>>>>>>>> So I just copy the hdfs-site.xml to >>>>>>>>>>>>>>~/.yarn/conf/hdfs-site.xml. >>>>>>>>>>>>>> Thank >>>>>>>>>>>>> you. >>>>>>>>>>>>>> >>>>>>>>>>>>>> Cheers, >>>>>>>>>>>>>> >>>>>>>>>>>>>> Fang, Yan >>>>>>>>>>>>>> [email protected] >>>>>>>>>>>>>> +1 (206) 849-4108 >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Tue, Mar 11, 2014 at 10:10 AM, Chris Riccomini >>>>>>>>>>>>>> <[email protected]>wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Hey Yan, >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Would you be up for contributing a tutorial page that >>>>>>>>>>>>>>> describes >>>>>>>>>>>>> this? >>>>>>>>>>>>>>> This >>>>>>>>>>>>>>> is really useful information. Our docs are just simple .md >>>>>>>>>>>>>>> files in the main code base. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Regarding step (3), is the hdfs-site.xml put into the conf >>>>>>>>>>>>>>> folder for the NM boxes, or on the client side (where >>>>>>>>>>>>>>> run-job.sh >>>>>>>>>> is run)? >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Cheers, >>>>>>>>>>>>>>> Chris >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> On 3/11/14 10:07 AM, "Yan Fang" <[email protected]> >>>>>>>>>>>>>>>>wrote: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Hi Sonali, >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> The way I make Samza run with HDFS is following: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> 1. include hdfs jar in Samza jar tar.gz. >>>>>>>>>>>>>>>> 2. you may also want to make sure the hadoop-common.jar >>>>>>>>>>>>>>>> has the same version as your hdfs jar. Otherwise, you may >>>>>>>>>>>>>>>> have configuration error popping out. >>>>>>>>>>>>>>>> 3. then put hdfs-site.xml to conf folder, the same folder >>>>>>>>>>>>>>>> as the yarn-site.xml 4. all other steps are not changed. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Hope this will help. Thank you. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Cheers, >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Fang, Yan >>>>>>>>>>>>>>>> [email protected] >>>>>>>>>>>>>>>> +1 (206) 849-4108 >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> On Tue, Mar 11, 2014 at 9:25 AM, Chris Riccomini >>>>>>>>>>>>>>>> <[email protected]>wrote: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Hey Sonali, >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> I believe that you need to make sure that the HDFS jar is >>>>>>>>>>>>>>>>> in your .tar.gz file, as you've said. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> If that doesn't work, you might need to define this >>>>>>>>>>>>>>>>> setting in core-site.xml on the machine you're running >>>>>>>>>> run-job.sh on: >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> <property> >>>>>>>>>>>>>>>>> <name>fs.hdfs.impl</name> >>>>>>>>>> <value>org.apache.hadoop.hdfs.DistributedFileSystem</value> >>>>>>>>>>>>>>>>> <description>The FileSystem for hdfs: >>>>>>>>>>>>>>>>> uris.</description> </property> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> You might also need to configure your NodeManagers to >>>>>>>>>>>>>>>>> have the HDFS >>>>>>>>>>>>>>> file >>>>>>>>>>>>>>>>> system impl as well. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> I've never run Samza with HDFS, so I'm guessing here. >>>>>>>>>>>>>>>>> Perhaps someone else on the list has been successful >>>>>>>>>>>>>>>>> with >>>>>>>>>> this? >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Cheers, >>>>>>>>>>>>>>>>> Chris >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> On 3/10/14 3:59 PM, "[email protected]" >>>>>>>>>>>>>>>>> <[email protected]> wrote: >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Hello, >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> I fixed this by starting from scratch with gradlew. But >>>>>>>>>>>>>>>>>> now when I >>>>>>>>>>>>>>> run >>>>>>>>>>>>>>>>> my >>>>>>>>>>>>>>>>>> job it throws this error: >>>>>>>>>>>>>>>>>> Exception in thread "main" java.io.IOException: No >>>>>>>>>>>>>>>>>> FileSystem for >>>>>>>>>>>>>>>>> scheme: >>>>>>>>>>>>>>>>>> hdfs >>>>>>>>>>>>>>>>>> at >>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSy >>>>>>>>>>>>>>>>>>ste >>>>>>>>>>>>>>>>>>m. >>>>>>>>>>>>>>>>>> jav >>>>>>>>>>>>>>>>>> a: >>>>>>>>>>>>>>>>>> 242 >>>>>>>>>>>>>>>>>> 1) >>>>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem >>>>>>>>>>>>>>>>. >>>>>>>>>>>>>>>> j >>>>>>>>>>>>>>>> a >>>>>>>>>>>>>>>> v >>>>>>>>>>>>>>>> a:2 >>>>>>>>>>>>>>>> 428 >>>>>>>>>>>>>>>> ) >>>>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java >>>>>>>>>>>>>>>>>: >>>>>>>>>>>>>>>>> 8 >>>>>>>>>>>>>>>>> 8 >>>>>>>>>>>>>>>>> ) >>>>>>>>>>>>>>>>>> at >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>>org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.j >>>>>>>>>>>>>ava >>>>>>>>>>>>>: >>>>>>>>>>>>>>>> 246 >>>>>>>>>>>>>>>> 7) >>>>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java: >>>>>>>>>>>>>>>>> 2 >>>>>>>>>>>>>>>>> 4 >>>>>>>>>>>>>>>>> 4 >>>>>>>>>>>>>>>>> 9) >>>>>>>>>>>>>>>>>> at >>>>>>>>>>>>> org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367) >>>>>>>>>>>>>>>>>> at >>>>>>>>>>>>> org.apache.hadoop.fs.Path.getFileSystem(Path.java:287) >>>>>>>>>>>>>>>>>> at >>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> org.apache.samza.job.yarn.ClientHelper.submitApplication >>>>>>>>>>>>>>>>>> (Cl >>>>>>>>>>>>>>>>>> i >>>>>>>>>>>>>>>>>> e >>>>>>>>>>>>>>>>>> n >>>>>>>>>>>>>>>>>> tHe >>>>>>>>>>>>>>>>>> lpe >>>>>>>>>>>>>>>>>> r. >>>>>>>>>>>>>>>>>> sc >>>>>>>>>>>>>>>>>> al >>>>>>>>>>>>>>>>>> a:111) >>>>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:55) >>>>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:48) >>>>>>>>>>>>>>>>>> at >>>>>>>>>>>>> org.apache.samza.job.JobRunner.run(JobRunner.scala:100) >>>>>>>>>>>>>>>>>> at >>>>>>>>>>>>> org.apache.samza.job.JobRunner$.main(JobRunner.scala:75) >>>>>>>>>>>>>>>>>> at >>>>>>>>>>>>>>>>>> org.apache.samza.job.JobRunner.main(JobRunner.scala) >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> I looked at the samza job tar.gz and it doesn't have a >>>>>>>>>>>>>>>>>> Hadoop-hdfs >>>>>>>>>>>>>>> jar. >>>>>>>>>>>>>>>>>> Is that why I get this error? >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>> Sonali >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> From: Parthasarathy, Sonali >>>>>>>>>>>>>>>>>> Sent: Monday, March 10, 2014 11:25 AM >>>>>>>>>>>>>>>>>> To: [email protected] >>>>>>>>>>>>>>>>>> Subject: Failed to package using mvn >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> When I tried to do a mvn clean package of my hello-samza >>>>>>>>>>>>>>>>>> project, I >>>>>>>>>>>>>>> get >>>>>>>>>>>>>>>>>> the following error. Has anyone seen this before? >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> [ERROR] Failed to execute goal on project >>>>>>>>>>>>>>>>>>samza-wikipedia: >>>>>>>>>>>>>>>>>> Could not resolve dependencies for project >>>>>>>>>>>>> samza:samza-wikipedia:jar:0.7.0: >>>>>>>>>>>>>>> Could >>>>>>>>>>>>>>>>>> not find artifact >>>>>>>>>>>>>>>>>> org.apache.samza:samza-kv_2.10:jar:0.7.0 >>>>>>>>>>>>>>>>>> in apache-releases >>>>>>>>>>>>>>> (https://repository.apache.org/content/groups/public) >>>>>>>>>>>>>>>>> -> >>>>>>>>>>>>>>>>>> [Help 1] >>>>>>>>>>>>>>>>>> [ERROR] >>>>>>>>>>>>>>>>>> [ERROR] To see the full stack trace of the errors, >>>>>>>>>>>>>>>>>> re-run Maven with >>>>>>>>>>>>>>>>> the >>>>>>>>>>>>>>>>>> -e switch. >>>>>>>>>>>>>>>>>> [ERROR] Re-run Maven using the -X switch to enable full >>>>>>>>>>>>>>>>>> debug >>>>>>>>>>>>>>> logging. >>>>>>>>>>>>>>>>>> [ERROR] >>>>>>>>>>>>>>>>>> [ERROR] For more information about the errors and >>>>>>>>>>>>>>>>>> possible >>>>>>>>>>>>>>> solutions, >>>>>>>>>>>>>>>>>> please read the following articles: >>>>>>>>>>>>>>>>>> [ERROR] [Help 1] >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> http://cwiki.apache.org/confluence/display/MAVEN/Dependency >>>>>>>>>>>>>>> Re >>>>>>>>>>>>>>> s >>>>>>>>>>>>>>> o >>>>>>>>>>>>>>> l >>>>>>>>>>>>>>> uti >>>>>>>>>>>>>>> onE >>>>>>>>>>>>>>> xce >>>>>>>>>>>>>>>>> p >>>>>>>>>>>>>>>>>> tion >>>>>>>>>>>>>>>>>> [ERROR] >>>>>>>>>>>>>>>>>> [ERROR] After correcting the problems, you can resume >>>>>>>>>>>>>>>>>> the build with >>>>>>>>>>>>>>>>> the >>>>>>>>>>>>>>>>>> command >>>>>>>>>>>>>>>>>> [ERROR] mvn <goals> -rf :samza-wikipedia >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>>>>>> Sonali >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> Sonali Parthasarathy >>>>>>>>>>>>>>>>>> R&D Developer, Data Insights Accenture Technology Labs >>>>>>>>>>>>>>>>>> 703-341-7432 >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> ________________________________ >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> This message is for the designated recipient only and >>>>>>>>>>>>>>>>>> may contain privileged, proprietary, or otherwise >>>>>>>>>>>>>>>>>> confidential >>>>>>>>>>>>> information. >>>>>>>>>>>>>>>>>> If >>>>>>>>>>>>>>> you >>>>>>>>>>>>>>>>>> have received it in error, please notify the sender >>>>>>>>>>>>>>>>>> immediately and delete the original. Any other use of >>>>>>>>>>>>>>>>>> the e-mail by you is >>>>>>>>>>>>>>> prohibited. >>>>>>>>>>>>>>>>>> Where allowed by local law, electronic communications >>>>>>>>>>>>>>>>>> with Accenture >>>>>>>>>>>>>>>>> and >>>>>>>>>>>>>>>>>> its affiliates, including e-mail and instant messaging >>>>>>>>>>>>>>>>>> (including content), may be scanned by our systems for >>>>>>>>>>>>>>>>>> the purposes of >>>>>>>>>>>>>>> information >>>>>>>>>>>>>>>>>> security and assessment of internal compliance with >>>>>>>>>>>>>>>>>> Accenture >>>>>>>>>>>>>>> policy. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> ________________________________________________________ >>>>>>>>>>>>>>>>>> __ >>>>>>>>>>>>>>>>>> _ >>>>>>>>>>>>>>>>>> _ >>>>>>>>>>>>>>>>>> _ >>>>>>>>>>>>>>>>>> ___ >>>>>>>>>>>>>>>>>> ___ >>>>>>>>>>>>>>>>>> ___ >>>>>>>>>>>>>>>>>> __ >>>>>>>>>>>>>>>>>> __ >>>>>>>>>>>>>>>>>> ____________ >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> www.accenture.com >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> ________________________________ >>>>>>>>>>>>> >>>>>>>>>>>>> This message is for the designated recipient only and may >>>>>>>>>>>>> contain privileged, proprietary, or otherwise confidential >>>>>>>>>>>>> information. If you have received it in error, please notify >>>>>>>>>>>>> the sender immediately and delete the original. Any other >>>>>>>>>>>>> use of the e-mail by you is prohibited. Where allowed by >>>>>>>>>>>>> local law, electronic communications with Accenture and its >>>>>>>>>>>>> affiliates, including e-mail and instant messaging (including >>>>>>>>>>>>> content), may be scanned by our systems for the purposes of >>>>>>>>>>>>> information security and assessment of internal compliance >>>>>>>>>>>>> with Accenture >>>>>>>>>> policy. >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> _____________________________________________________________ >>>>>>>>>>>>> ___ >>>>>>>>>>>>> _ >>>>>>>>>>>>> _ >>>>>>>>>>>>> _ >>>>>>>>>>>>> ___ >>>>>>>>>>>>> ___ >>>>>>>>>>>>> _____________ >>>>>>>>>>>>> >>>>>>>>>>>>> www.accenture.com >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> ________________________________ >>>>>>>>>>> >>>>>>>>>>> This message is for the designated recipient only and may >>>>>>>>>>> contain privileged, proprietary, or otherwise confidential >>>>>>>>>>> information. If you have received it in error, please notify >>>>>>>>>>> the sender immediately and delete the original. Any other use >>>>>>>>>>> of the e-mail by you is >>>>>>>>>> prohibited. >>>>>>>>>>> Where allowed by local law, electronic communications with >>>>>>>>>>> Accenture and its affiliates, including e-mail and instant >>>>>>>>>>> messaging (including content), may be scanned by our systems >>>>>>>>>>> for the purposes of information security and assessment of >>>>>>>>>>> internal compliance with >>>>>>>>>> Accenture policy. >>>>>>>>>>> _______________________________________________________________ >>>>>>>>>>> ___ >>>>>>>>>>> _ >>>>>>>>>>> _ >>>>>>>>>>> _ >>>>>>>>>>> _____ >>>>>>>>>>> ____________ >>>>>>>>>>> >>>>>>>>>>> www.accenture.com >>>>>>>>> >>>>>>>>> ________________________________ >>>>>>>>> >>>>>>>>> This message is for the designated recipient only and may contain >>>>>>>>>privileged, proprietary, or otherwise confidential information. If >>>>>>>>>you have received it in error, please notify the sender >>>>>>>>>immediately and delete the original. Any other use of the e-mail >>>>>>>>>by you is prohibited. >>>>>>>>> Where allowed by local law, electronic communications with >>>>>>>>>Accenture and its affiliates, including e-mail and instant >>>>>>>>>messaging (including content), may be scanned by our systems for >>>>>>>>>the purposes of information security and assessment of internal >>>>>>>>>compliance with Accenture policy. >>>>>>>>> >>>>>>>>>__________________________________________________________________ >>>>>>>>>___ >>>>>>>>> _ >>>>>>>>> _ >>>>>>>>> ___ >>>>>>>>> ____________ >>>>>>>>> >>>>>>>>> www.accenture.com >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> ________________________________ >>>>>>>> >>>>>>>> This message is for the designated recipient only and may contain >>>>>>>>privileged, proprietary, or otherwise confidential information. If >>>>>>>>you have received it in error, please notify the sender >>>>>>>>immediately and delete the original. Any other use of the e-mail >>>>>>>>by you is prohibited. >>>>>>>> Where allowed by local law, electronic communications with >>>>>>>>Accenture and its affiliates, including e-mail and instant >>>>>>>>messaging (including content), may be scanned by our systems for >>>>>>>>the purposes of information security and assessment of internal >>>>>>>>compliance with Accenture policy. >>>>>>>> >>>>>>>>___________________________________________________________________ >>>>>>>>___ >>>>>>>> _ >>>>>>>> ___ >>>>>>>> ____________ >>>>>>>> >>>>>>>> www.accenture.com >>>>>>> >>>>>>> >>>>>>> >>>>>>> ________________________________ >>>>>>> >>>>>>> This message is for the designated recipient only and may contain >>>>>>>privileged, proprietary, or otherwise confidential information. If >>>>>>>you have received it in error, please notify the sender immediately >>>>>>>and delete the original. Any other use of the e-mail by you is >>>>>>>prohibited. >>>>>>> Where allowed by local law, electronic communications with >>>>>>>Accenture and its affiliates, including e-mail and instant >>>>>>>messaging (including content), may be scanned by our systems for >>>>>>>the purposes of information security and assessment of internal >>>>>>>compliance with Accenture policy. >>>>>>> >>>>>>>____________________________________________________________________ >>>>>>>___ >>>>>>> ___ >>>>>>> ____________ >>>>>>> >>>>>>> www.accenture.com >>>>>> >>>>>> >>>>>> >>>>>> ________________________________ >>>>>> >>>>>> This message is for the designated recipient only and may contain >>>>>>privileged, proprietary, or otherwise confidential information. If >>>>>>you have received it in error, please notify the sender immediately >>>>>>and delete the original. Any other use of the e-mail by you is >>>>>>prohibited. >>>>>> Where allowed by local law, electronic communications with Accenture >>>>>>and its affiliates, including e-mail and instant messaging >>>>>>(including content), may be scanned by our systems for the purposes >>>>>>of information security and assessment of internal compliance with >>>>>>Accenture policy. >>>>>> >>>>>>_____________________________________________________________________ >>>>>>___ >>>>>>__ >>>>>> ____________ >>>>>> >>>>>> www.accenture.com >>>>> >>> >>> >>> >>>________________________________ >>> >>>This message is for the designated recipient only and may contain >>>privileged, proprietary, or otherwise confidential information. If you >>>have received it in error, please notify the sender immediately and >>>delete the original. Any other use of the e-mail by you is prohibited. >>>Where allowed by local law, electronic communications with Accenture and >>>its affiliates, including e-mail and instant messaging (including >>>content), may be scanned by our systems for the purposes of information >>>security and assessment of internal compliance with Accenture policy. >>>________________________________________________________________________ >>>_ >>>_ >>>____________ >>> >>>www.accenture.com >>> >> > > > >________________________________ > >This message is for the designated recipient only and may contain >privileged, proprietary, or otherwise confidential information. If you >have received it in error, please notify the sender immediately and >delete the original. Any other use of the e-mail by you is prohibited. >Where allowed by local law, electronic communications with Accenture and >its affiliates, including e-mail and instant messaging (including >content), may be scanned by our systems for the purposes of information >security and assessment of internal compliance with Accenture policy. >__________________________________________________________________________ >____________ > >www.accenture.com >
