I guess Sonali has the problem is because his NMs do not read the YARN_HOME 
variable. That may be because the NM machine does not have YARN_HOME set when 
the NM starts. 

Check this https://issues.apache.org/jira/browse/SAMZA-182

Thanks,

Yan Fang

> On Mar 12, 2014, at 6:14 PM, Chris Riccomini <[email protected]> wrote:
> 
> Hey Sonali,
> 
> I am unfamiliar with the start-yarn.sh. Looking at:
> 
> 
> https://svn.apache.org/viewvc/hadoop/common/trunk/hadoop-yarn-project/hadoo
> p-yarn/bin/stop-yarn.sh?revision=1370666&view=markup
> 
> What version of YARN are you using?
> 
> Cheers,
> Chris
> 
> On 3/12/14 5:56 PM, "[email protected]"
> <[email protected]> wrote:
> 
>> Hey Chris,
>> 
>> Yes, I have YARN_HOME set in all the NMs pointing to the right
>> directories. I also made sure the yarn-site.xml file has the hostname set.
>> 
>> I start yarn using start.yarn.sh in the RM and that automatically starts
>> the NMs on the slave nodes. Is that the right way to do it?
>> 
>> -----Original Message-----
>> From: Chris Riccomini [mailto:[email protected]]
>> Sent: Wednesday, March 12, 2014 5:52 PM
>> To: [email protected]
>> Subject: Re: Failed to package using mvn
>> 
>> Hey Sonali,
>> 
>> OK, so we've validated that the NMs are able to connect, which means they
>> can see the yarn-site.xml.
>> 
>> How are you starting your NMs? Are you running:
>> 
>> export YARN_HOME=/path/to/yarn/home
>> 
>> In the CLI before starting the NM?
>> 
>> For reference, we run:
>> 
>> export YARN_HOME=/path/to/our/yarn-home
>> export YARN_CONF_DIR=$YARN_HOME/conf
>> 
>> bin/yarn nodemanager
>> 
>> With YARN_HOME pointing to a directory that has a subdirectory called
>> "conf" in it, which has a yarn-site.xml in it:
>> 
>> /path/to/our/yarn-home/conf/yarn-site.xml
>> 
>> This yarn-site.xml has yarn.resourcemanager.hostname set to the IP (or
>> hostname) of the resource manager:
>> 
>> <property>
>> <name>yarn.resourcemanager.hostname</name>
>> <value>123.456.789.123</value>
>> </property>
>> 
>> 
>> Cheers,
>> Chris
>> 
>> On 3/12/14 5:33 PM, "[email protected]"
>> <[email protected]> wrote:
>> 
>>> I see two active nodes (I have 2 NMs running)
>>> 
>>> -----Original Message-----
>>> From: Chris Riccomini [mailto:[email protected]]
>>> Sent: Wednesday, March 12, 2014 5:24 PM
>>> To: [email protected]
>>> Subject: Re: Failed to package using mvn
>>> 
>>> Hey Sonali,
>>> 
>>> Can you go to your ResourceManager's UI, and tell me how many active
>>> nodes you see? This should be under the "active nodes" heading.
>>> 
>>> It sounds like the SamzaAppMaster is not getting the resource manager
>>> host/port from the yarn-site.xml. Usually this is due to not exporting
>>> YARN_HOME on the NodeManager before starting it.
>>> 
>>> Cheers,
>>> Chris
>>> 
>>> On 3/12/14 5:21 PM, "[email protected]"
>>> <[email protected]> wrote:
>>> 
>>>> Okay so I was able to submit the job:
>>>> 
>>>> In the nodemanager I get this error:  Specifically it's trying to
>>>> connect to 0.0.0.0/8032 instead of the IP I have specified in the
>>>> yarn-site.xml file
>>>> 
>>>> 2014-03-12 17:04:47 SamzaAppMaster$ [INFO] got container id:
>>>> container_1391637982288_0033_01_000001
>>>> 2014-03-12 17:04:47 SamzaAppMaster$ [INFO] got app attempt id:
>>>> appattempt_1391637982288_0033_000001
>>>> 2014-03-12 17:04:47 SamzaAppMaster$ [INFO] got node manager host:
>>>> svdpdac001.techlabs.accenture.com
>>>> 2014-03-12 17:04:47 SamzaAppMaster$ [INFO] got node manager port:
>>>> 38218
>>>> 2014-03-12 17:04:47 SamzaAppMaster$ [INFO] got node manager http port:
>>>> 8042
>>>> 2014-03-12 17:04:47 SamzaAppMaster$ [INFO] got config:
>>>> {task.inputs=wikipedia.#en.wikipedia,wikipedia.#en.wiktionary,wikipedi
>>>> a .#e n.wikinews, systems.wikipedia.host=irc.wikimedia.org,
>>>> systems.kafka.producer.batch.num.messages=1,
>>>> job.factory.class=org.apache.samza.job.yarn.YarnJobFactory,
>>>> systems.wikipedia.port=6667,
>>>> systems.kafka.producer.producer.type=sync,
>>>> job.name=wikipedia-feed,
>>>> systems.kafka.consumer.zookeeper.connect=svdpdac013.techlabs.accenture.
>>>> com :2181/, systems.kafka.samza.msg.serde=json,
>>>> serializers.registry.json.class=org.apache.samza.serializers.JsonSerde
>>>> F
>>>> act ory,
>>>> task.class=samza.examples.wikipedia.task.WikipediaFeedStreamTask,
>>>> yarn.package.path=hdfs://10.1.174.85:9000/samza-job-package-0.7.0-dist.
>>>> tar
>>>> .gz,
>>>> systems.wikipedia.samza.factory=samza.examples.wikipedia.system.Wikipe
>>>> d
>>>> iaS
>>>> ystemFactory,
>>>> systems.kafka.samza.factory=org.apache.samza.system.kafka.KafkaSystemF
>>>> a
>>>> cto
>>>> ry,
>>>> systems.kafka.producer.metadata.broker.list=svdpdac001.techlabs.accent
>>>> ure
>>>> .
>>>> com:6667,svdpdac015.techlabs.accenture.com:6667}
>>>> 2014-03-12 17:04:48 ClientHelper [INFO] trying to connect to RM
>>>> 0.0.0.0:8032
>>>> 2014-03-12 17:04:48 NativeCodeLoader [WARN] Unable to load
>>>> native-hadoop library for your platform... using builtin-java classes
>>>> where applicable
>>>> 2014-03-12 17:04:48 RMProxy [INFO] Connecting to ResourceManager at
>>>> /0.0.0.0:8032
>>>> 
>>>> -----Original Message-----
>>>> From: Chris Riccomini [mailto:[email protected]]
>>>> Sent: Wednesday, March 12, 2014 4:48 PM
>>>> To: [email protected]
>>>> Subject: Re: Failed to package using mvn
>>>> 
>>>> Hey Sonali,
>>>> 
>>>> You need to specify a valid HDFS uri. Usually something like:
>>>> 
>>>> hdfs://<hdfs name node ip>:<hdfs name node port>/path/to/tgz
>>>> 
>>>> Right now, Hadoop is trying to use the package name as the HDFS host.
>>>> 
>>>> Cheers,
>>>> Chris
>>>> 
>>>> On 3/12/14 4:45 PM, "[email protected]"
>>>> <[email protected]> wrote:
>>>> 
>>>>> I did and I can now see the hadoop-hdfs jar in /deploy/samza/lib
>>>>> folder.
>>>>> 
>>>>> I do get a different error now.
>>>>> 
>>>>> I uploaded the samza-job to hdfs and it resides on
>>>>> hdfs://samza-job-package-0.7.0-dist.tar.gz
>>>>> 
>>>>> But when I run the job I get this exception:
>>>>> 
>>>>> Exception in thread "main" java.lang.IllegalArgumentException:
>>>>> java.net.UnknownHostException: samza-job-package-0.7.0-dist.tar.gz
>>>>>      at
>>>>> org.apache.hadoop.security.SecurityUtil.buildTokenService(SecurityUtil.
>>>>> jav
>>>>> a:418)
>>>>>      at
>>>>> org.apache.hadoop.hdfs.NameNodeProxies.createNonHAProxy(NameNodeProxi
>>>>> e
>>>>> s
>>>>> .ja
>>>>> va:231)
>>>>>      at
>>>>> org.apache.hadoop.hdfs.NameNodeProxies.createProxy(NameNodeProxies.ja
>>>>> v
>>>>> a
>>>>> :13
>>>>> 9)
>>>>>      at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:510)
>>>>>      at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:453)
>>>>>      at
>>>>> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFi
>>>>> l
>>>>> e
>>>>> Sys
>>>>> tem.java:136)
>>>>>      at
>>>>> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:2433)
>>>>>      at
>>>>> org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:88)
>>>>>      at
>>>>> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2467)
>>>>>      at
>>>>> org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2449)
>>>>>      at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367)
>>>>>      at org.apache.hadoop.fs.Path.getFileSystem(Path.java:287)
>>>>>      at
>>>>> org.apache.samza.job.yarn.ClientHelper.submitApplication(ClientHelper.
>>>>> s
>>>>> cal
>>>>> a:111)
>>>>>      at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:55)
>>>>>      at org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:48)
>>>>>      at org.apache.samza.job.JobRunner.run(JobRunner.scala:100)
>>>>>      at org.apache.samza.job.JobRunner$.main(JobRunner.scala:75)
>>>>>      at org.apache.samza.job.JobRunner.main(JobRunner.scala)
>>>>> Caused by: java.net.UnknownHostException:
>>>>> samza-job-package-0.7.0-dist.tar.gz
>>>>>      ... 18 more
>>>>> 
>>>>> 
>>>>> -----Original Message-----
>>>>> From: Yan Fang [mailto:[email protected]]
>>>>> Sent: Wednesday, March 12, 2014 4:20 PM
>>>>> To: [email protected]
>>>>> Subject: Re: Failed to package using mvn
>>>>> 
>>>>> Hi Sonali,
>>>>> 
>>>>> One tip you may miss:
>>>>> 
>>>>> If you had already run
>>>>> 
>>>>> tar -xvf
>>>>> ./samza-job-package/target/samza-job-package-0.7.0-dist.tar.gz
>>>>> -C deploy/samza
>>>>> 
>>>>> before you bundled the jar file to tar.gz. Please also remember to
>>>>> put the hdfs jar file to the deploy/samza/lib.
>>>>> 
>>>>> Let me know if you miss this step.
>>>>> 
>>>>> Thanks,
>>>>> 
>>>>> Fang, Yan
>>>>> [email protected]
>>>>> +1 (206) 849-4108
>>>>> 
>>>>> 
>>>>> On Wed, Mar 12, 2014 at 4:10 PM, Chris Riccomini
>>>>> <[email protected]>wrote:
>>>>> 
>>>>>> Hey Sonali,
>>>>>> 
>>>>>> Yan has made a step-by-step tutorial for this. Could you confirm
>>>>>> that you've followed the instructions, and it's still not working?
>>>>>> 
>>>>>> https://issues.apache.org/jira/browse/SAMZA-181
>>>>>> 
>>>>>> Cheers,
>>>>>> Chris
>>>>>> 
>>>>>> On 3/12/14 3:12 PM, "[email protected]"
>>>>>> <[email protected]> wrote:
>>>>>> 
>>>>>>> So sigh! I had some Kafka issues in-between. That's fixed now.
>>>>>>> 
>>>>>>> As suggested,
>>>>>>> 
>>>>>>> 1. I made sure the hadoop-hdfs-2.2.0.jar is bundled with the samza
>>>>>>> job tar.gz.
>>>>>>> 2. I added the configuration to implement hdfs in the
>>>>>>> hdfs-site.xml files both on the NMs and in the /conf directory for
>>>>>>> samza
>>>>>>> 
>>>>>>> I still get the No Filesystem for scheme :hdfs error.
>>>>>>> 
>>>>>>> Is there anything else im missing?
>>>>>>> Thanks,
>>>>>>> Sonali
>>>>>>> 
>>>>>>> 
>>>>>>> -----Original Message-----
>>>>>>> From: Chris Riccomini [mailto:[email protected]]
>>>>>>> Sent: Tuesday, March 11, 2014 8:27 PM
>>>>>>> To: [email protected]
>>>>>>> Subject: Re: Failed to package using mvn
>>>>>>> 
>>>>>>> Hey Yan,
>>>>>>> 
>>>>>>> This looks great! I added a few requests to the JIRA, if you have
>>>>>> time.
>>>>>>> 
>>>>>>> Cheers,
>>>>>>> Chris
>>>>>>> 
>>>>>>>> On 3/11/14 7:20 PM, "Yan Fang" <[email protected]> wrote:
>>>>>>>> 
>>>>>>>> Hi Chris,
>>>>>>>> 
>>>>>>>> Has opened an issue
>>>>>>>> SAMZA-181<https://issues.apache.org/jira/browse/SAMZA-181>and
>>>>>>>> also uploaded the patch. Let me know if there is something wrong
>>>>>>>> in my tutorial. Thank you!
>>>>>>>> 
>>>>>>>> Cheers,
>>>>>>>> 
>>>>>>>> Fang, Yan
>>>>>>>> [email protected]
>>>>>>>> +1 (206) 849-4108
>>>>>>>> 
>>>>>>>> 
>>>>>>>> On Tue, Mar 11, 2014 at 10:40 AM,
>>>>>>>> <[email protected]>wrote:
>>>>>>>> 
>>>>>>>>> Thanks Chris, Yan,
>>>>>>>>> 
>>>>>>>>> Let me try that.
>>>>>>>>> 
>>>>>>>>> -----Original Message-----
>>>>>>>>> From: Chris Riccomini [mailto:[email protected]]
>>>>>>>>> Sent: Tuesday, March 11, 2014 10:22 AM
>>>>>>>>> To: [email protected]
>>>>>>>>> Subject: Re: Failed to package using mvn
>>>>>>>>> 
>>>>>>>>> Hey Yan,
>>>>>>>>> 
>>>>>>>>> Awesome!The location where you can add your .md is here:
>>>>>>>>> 
>>>>>>>>> docs/learn/tutorials/0.7.0/
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> Here's a link to the code tree:
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> https://git-wip-us.apache.org/repos/asf?p=incubator-samza.git;a=
>>>>>>>>> t
>>>>>>>>> r
>>>>>>>>> e
>>>>>>>>> e;f
>>>>>>>>> =do
>>>>>>>>> cs
>>>>>>>>> 
>>>>>>>>> /learn/tutorials/0.7.0;h=ef117f4066f14a00f50f0f6fca1790313044831
>>>>>>>>> 2
>>>>>>>>> ;
>>>>>>>>> h
>>>>>>>>> b=H
>>>>>>>>> EAD
>>>>>>>>> 
>>>>>>>>> You can get the code here:
>>>>>>>>> 
>>>>>>>>> git clone
>>>>>>>>> http://git-wip-us.apache.org/repos/asf/incubator-samza.git
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> Once you write the .md, just throw it up on a JIRA, and one of
>>>>>>>>> us can merge it in.
>>>>>>>>> 
>>>>>>>>> Re: hdfs-site.xml, ah ha, that's what I figured. This is good
>>>>>>>>> to
>>>>>> know.
>>>>>>>>> So
>>>>>>>>> you just copy your hdfs-site.xml from your NodeManager's conf
>>>>>>>>> directory  into your local hdfs-site.xml.
>>>>>>>>> 
>>>>>>>>> Cheers,
>>>>>>>>> Chris
>>>>>>>>> 
>>>>>>>>>> On 3/11/14 10:16 AM, "Yan Fang" <[email protected]> wrote:
>>>>>>>>>> 
>>>>>>>>>> Hi Chris,
>>>>>>>>>> 
>>>>>>>>>> Sure. I just do not know how/where to contribute this
>>>>>>>>>> page...*_*
>>>>>>>>>> 
>>>>>>>>>> Oh, I mean the same this as you mentioned in the *Cluster
>>>>>>>>>> Installation*thread:
>>>>>>>>>> 
>>>>>>>>>> *"2. Get a copy of one of your NM's yarn-site.xml and put it
>>>>>>>>>> somewhere
>>>>>>>>>> on*
>>>>>>>>>> 
>>>>>>>>>> *your desktop (I usually use ~/.yarn/conf/yarn-site.xml). Note
>>>>>>>>>> that there'sa "conf" directory there. This is mandatory."*
>>>>>>>>>> 
>>>>>>>>>> So I just copy the hdfs-site.xml to ~/.yarn/conf/hdfs-site.xml.
>>>>>>>>>> Thank
>>>>>>>>> you.
>>>>>>>>>> 
>>>>>>>>>> Cheers,
>>>>>>>>>> 
>>>>>>>>>> Fang, Yan
>>>>>>>>>> [email protected]
>>>>>>>>>> +1 (206) 849-4108
>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>>> On Tue, Mar 11, 2014 at 10:10 AM, Chris Riccomini
>>>>>>>>>> <[email protected]>wrote:
>>>>>>>>>> 
>>>>>>>>>>> Hey Yan,
>>>>>>>>>>> 
>>>>>>>>>>> Would you be up for contributing a tutorial page that
>>>>>>>>>>> describes
>>>>>>>>> this?
>>>>>>>>>>> This
>>>>>>>>>>> is really useful information. Our docs are just simple .md
>>>>>>>>>>> files in the  main code base.
>>>>>>>>>>> 
>>>>>>>>>>> Regarding step (3), is the hdfs-site.xml put into the conf
>>>>>>>>>>> folder for the  NM boxes, or on the client side (where
>>>>>>>>>>> run-job.sh
>>>>>> is run)?
>>>>>>>>>>> 
>>>>>>>>>>> Cheers,
>>>>>>>>>>> Chris
>>>>>>>>>>> 
>>>>>>>>>>>> On 3/11/14 10:07 AM, "Yan Fang" <[email protected]> wrote:
>>>>>>>>>>>> 
>>>>>>>>>>>> Hi Sonali,
>>>>>>>>>>>> 
>>>>>>>>>>>> The way I make Samza run with HDFS is following:
>>>>>>>>>>>> 
>>>>>>>>>>>> 1. include hdfs jar in Samza jar tar.gz.
>>>>>>>>>>>> 2. you may also want to make sure the hadoop-common.jar has
>>>>>>>>>>>> the same version as your hdfs jar. Otherwise, you may have
>>>>>>>>>>>> configuration error popping out.
>>>>>>>>>>>> 3. then put hdfs-site.xml to conf folder, the same folder
>>>>>>>>>>>> as the yarn-site.xml 4. all other steps are not changed.
>>>>>>>>>>>> 
>>>>>>>>>>>> Hope this will help. Thank you.
>>>>>>>>>>>> 
>>>>>>>>>>>> Cheers,
>>>>>>>>>>>> 
>>>>>>>>>>>> Fang, Yan
>>>>>>>>>>>> [email protected]
>>>>>>>>>>>> +1 (206) 849-4108
>>>>>>>>>>>> 
>>>>>>>>>>>> 
>>>>>>>>>>>> On Tue, Mar 11, 2014 at 9:25 AM, Chris Riccomini
>>>>>>>>>>>> <[email protected]>wrote:
>>>>>>>>>>>> 
>>>>>>>>>>>>> Hey Sonali,
>>>>>>>>>>>>> 
>>>>>>>>>>>>> I believe that you need to make sure that the HDFS jar is
>>>>>>>>>>>>> in your .tar.gz  file, as you've said.
>>>>>>>>>>>>> 
>>>>>>>>>>>>> If that doesn't work, you might need to define this
>>>>>>>>>>>>> setting in core-site.xml on the machine you're running
>>>>>> run-job.sh on:
>>>>>>>>>>>>> 
>>>>>>>>>>>>> <property>
>>>>>>>>>>>>> <name>fs.hdfs.impl</name>
>>>>>> <value>org.apache.hadoop.hdfs.DistributedFileSystem</value>
>>>>>>>>>>>>> <description>The FileSystem for hdfs:
>>>>>>>>>>>>> uris.</description> </property>
>>>>>>>>>>>>> 
>>>>>>>>>>>>> 
>>>>>>>>>>>>> You might also need to configure your NodeManagers to
>>>>>>>>>>>>> have the HDFS
>>>>>>>>>>> file
>>>>>>>>>>>>> system impl as well.
>>>>>>>>>>>>> 
>>>>>>>>>>>>> I've never run Samza with HDFS, so I'm guessing here.
>>>>>>>>>>>>> Perhaps someone else  on the list has been successful with
>>>>>> this?
>>>>>>>>>>>>> 
>>>>>>>>>>>>> Cheers,
>>>>>>>>>>>>> Chris
>>>>>>>>>>>>> 
>>>>>>>>>>>>> On 3/10/14 3:59 PM, "[email protected]"
>>>>>>>>>>>>> <[email protected]> wrote:
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Hello,
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> I fixed this by starting from scratch with gradlew. But
>>>>>>>>>>>>>> now when I
>>>>>>>>>>> run
>>>>>>>>>>>>> my
>>>>>>>>>>>>>> job it throws this error:
>>>>>>>>>>>>>> Exception in thread "main" java.io.IOException: No
>>>>>>>>>>>>>> FileSystem for
>>>>>>>>>>>>> scheme:
>>>>>>>>>>>>>> hdfs
>>>>>>>>>>>>>>      at
>>>>>> 
>>>>>>>>>>>>>> org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.
>>>>>>>>>>>>>> jav
>>>>>>>>>>>>>> a:
>>>>>>>>>>>>>> 242
>>>>>>>>>>>>>> 1)
>>>>>>>>>>>>>>      at
>>>>>>>>>>>> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.
>>>>>>>>>>>> j
>>>>>>>>>>>> a
>>>>>>>>>>>> v
>>>>>>>>>>>> a:2
>>>>>>>>>>>> 428
>>>>>>>>>>>> )
>>>>>>>>>>>>>>      at
>>>>>>>>>>>>> org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:
>>>>>>>>>>>>> 8
>>>>>>>>>>>>> 8
>>>>>>>>>>>>> )
>>>>>>>>>>>>>>      at
>>>>>>>>> 
>>>>>>>>> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:
>>>>>>>>>>>> 246
>>>>>>>>>>>> 7)
>>>>>>>>>>>>>>      at
>>>>>>>>>>>>> org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:
>>>>>>>>>>>>> 2
>>>>>>>>>>>>> 4
>>>>>>>>>>>>> 4
>>>>>>>>>>>>> 9)
>>>>>>>>>>>>>>      at
>>>>>>>>> org.apache.hadoop.fs.FileSystem.get(FileSystem.java:367)
>>>>>>>>>>>>>>      at
>>>>>>>>> org.apache.hadoop.fs.Path.getFileSystem(Path.java:287)
>>>>>>>>>>>>>>      at
>>>>>>>>> 
>>>>>>>>>>>>>> org.apache.samza.job.yarn.ClientHelper.submitApplication(Cl
>>>>>>>>>>>>>> i
>>>>>>>>>>>>>> e
>>>>>>>>>>>>>> n
>>>>>>>>>>>>>> tHe
>>>>>>>>>>>>>> lpe
>>>>>>>>>>>>>> r.
>>>>>>>>>>>>>> sc
>>>>>>>>>>>>>> al
>>>>>>>>>>>>>> a:111)
>>>>>>>>>>>>>>      at
>>>>>>>>>>> org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:55)
>>>>>>>>>>>>>>      at
>>>>>>>>>>> org.apache.samza.job.yarn.YarnJob.submit(YarnJob.scala:48)
>>>>>>>>>>>>>>      at
>>>>>>>>> org.apache.samza.job.JobRunner.run(JobRunner.scala:100)
>>>>>>>>>>>>>>      at
>>>>>>>>> org.apache.samza.job.JobRunner$.main(JobRunner.scala:75)
>>>>>>>>>>>>>>      at
>>>>>>>>>>>>>> org.apache.samza.job.JobRunner.main(JobRunner.scala)
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> I looked at the samza job tar.gz and it doesn't have a
>>>>>>>>>>>>>> Hadoop-hdfs
>>>>>>>>>>> jar.
>>>>>>>>>>>>>> Is that why I get this error?
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>> Sonali
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> From: Parthasarathy, Sonali
>>>>>>>>>>>>>> Sent: Monday, March 10, 2014 11:25 AM
>>>>>>>>>>>>>> To: [email protected]
>>>>>>>>>>>>>> Subject: Failed to package using mvn
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Hi,
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> When I tried to do a mvn clean package of my hello-samza
>>>>>>>>>>>>>> project, I
>>>>>>>>>>> get
>>>>>>>>>>>>>> the following error. Has anyone seen this before?
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> [ERROR] Failed to execute goal on project samza-wikipedia:
>>>>>>>>>>>>>> Could not resolve dependencies for project
>>>>>>>>> samza:samza-wikipedia:jar:0.7.0:
>>>>>>>>>>> Could
>>>>>>>>>>>>>> not find artifact
>>>>>>>>>>>>>> org.apache.samza:samza-kv_2.10:jar:0.7.0
>>>>>>>>>>>>>> in apache-releases
>>>>>>>>>>> (https://repository.apache.org/content/groups/public)
>>>>>>>>>>>>> ->
>>>>>>>>>>>>>> [Help 1]
>>>>>>>>>>>>>> [ERROR]
>>>>>>>>>>>>>> [ERROR] To see the full stack trace of the errors,
>>>>>>>>>>>>>> re-run Maven with
>>>>>>>>>>>>> the
>>>>>>>>>>>>>> -e switch.
>>>>>>>>>>>>>> [ERROR] Re-run Maven using the -X switch to enable full
>>>>>>>>>>>>>> debug
>>>>>>>>>>> logging.
>>>>>>>>>>>>>> [ERROR]
>>>>>>>>>>>>>> [ERROR] For more information about the errors and
>>>>>>>>>>>>>> possible
>>>>>>>>>>> solutions,
>>>>>>>>>>>>>> please read the following articles:
>>>>>>>>>>>>>> [ERROR] [Help 1]
>>>>>>>>>>> 
>>>>>>>>>>> http://cwiki.apache.org/confluence/display/MAVEN/DependencyRe
>>>>>>>>>>> s
>>>>>>>>>>> o
>>>>>>>>>>> l
>>>>>>>>>>> uti
>>>>>>>>>>> onE
>>>>>>>>>>> xce
>>>>>>>>>>>>> p
>>>>>>>>>>>>>> tion
>>>>>>>>>>>>>> [ERROR]
>>>>>>>>>>>>>> [ERROR] After correcting the problems, you can resume
>>>>>>>>>>>>>> the build with
>>>>>>>>>>>>> the
>>>>>>>>>>>>>> command
>>>>>>>>>>>>>> [ERROR]   mvn <goals> -rf :samza-wikipedia
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>> Sonali
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Sonali Parthasarathy
>>>>>>>>>>>>>> R&D Developer, Data Insights Accenture Technology Labs
>>>>>>>>>>>>>> 703-341-7432
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> ________________________________
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> This message is for the designated recipient only and
>>>>>>>>>>>>>> may contain privileged, proprietary, or otherwise
>>>>>>>>>>>>>> confidential
>>>>>>>>> information.
>>>>>>>>>>>>>> If
>>>>>>>>>>> you
>>>>>>>>>>>>>> have received it in error, please notify the sender
>>>>>>>>>>>>>> immediately and delete the original. Any other use of
>>>>>>>>>>>>>> the e-mail by you is
>>>>>>>>>>> prohibited.
>>>>>>>>>>>>>> Where allowed by local law, electronic communications
>>>>>>>>>>>>>> with Accenture
>>>>>>>>>>>>> and
>>>>>>>>>>>>>> its affiliates, including e-mail and instant messaging
>>>>>>>>>>>>>> (including content), may be scanned by our systems for
>>>>>>>>>>>>>> the purposes of
>>>>>>>>>>> information
>>>>>>>>>>>>>> security and assessment of internal compliance with
>>>>>>>>>>>>>> Accenture
>>>>>>>>>>> policy.
>>>>>>>>>>> 
>>>>>>>>>>>>>> __________________________________________________________
>>>>>>>>>>>>>> _
>>>>>>>>>>>>>> _
>>>>>>>>>>>>>> _
>>>>>>>>>>>>>> ___
>>>>>>>>>>>>>> ___
>>>>>>>>>>>>>> ___
>>>>>>>>>>>>>> __
>>>>>>>>>>>>>> __
>>>>>>>>>>>>>> ____________
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> www.accenture.com
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> ________________________________
>>>>>>>>> 
>>>>>>>>> This message is for the designated recipient only and may
>>>>>>>>> contain privileged, proprietary, or otherwise confidential
>>>>>>>>> information. If you have  received it in error, please notify
>>>>>>>>> the sender immediately and delete the  original. Any other use
>>>>>>>>> of the e-mail by you is prohibited. Where allowed  by local law,
>>>>>>>>> electronic communications with Accenture and its affiliates,
>>>>>>>>> including e-mail and instant messaging (including content), may
>>>>>>>>> be scanned by our systems for the purposes of information
>>>>>>>>> security and assessment of internal compliance with Accenture
>>>>>> policy.
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> ________________________________________________________________
>>>>>>>>> _
>>>>>>>>> _
>>>>>>>>> _
>>>>>>>>> ___
>>>>>>>>> ___
>>>>>>>>> _____________
>>>>>>>>> 
>>>>>>>>> www.accenture.com
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> ________________________________
>>>>>>> 
>>>>>>> This message is for the designated recipient only and may contain
>>>>>>> privileged, proprietary, or otherwise confidential information. If
>>>>>>> you have received it in error, please notify the sender
>>>>>>> immediately and delete the original. Any other use of the e-mail
>>>>>>> by you is
>>>>>> prohibited.
>>>>>>> Where allowed by local law, electronic communications with
>>>>>>> Accenture and its affiliates, including e-mail and instant
>>>>>>> messaging (including content), may be scanned by our systems for
>>>>>>> the purposes of information security and assessment of internal
>>>>>>> compliance with
>>>>>> Accenture policy.
>>>>>>> __________________________________________________________________
>>>>>>> _
>>>>>>> _
>>>>>>> _
>>>>>>> _____
>>>>>>> ____________
>>>>>>> 
>>>>>>> www.accenture.com
>>>>> 
>>>>> ________________________________
>>>>> 
>>>>> This message is for the designated recipient only and may contain
>>>>> privileged, proprietary, or otherwise confidential information. If
>>>>> you have received it in error, please notify the sender immediately
>>>>> and delete the original. Any other use of the e-mail by you is
>>>>> prohibited.
>>>>> Where allowed by local law, electronic communications with Accenture
>>>>> and its affiliates, including e-mail and instant messaging (including
>>>>> content), may be scanned by our systems for the purposes of
>>>>> information security and assessment of internal compliance with
>>>>> Accenture policy.
>>>>> _____________________________________________________________________
>>>>> _
>>>>> _
>>>>> ___
>>>>> ____________
>>>>> 
>>>>> www.accenture.com
>>>> 
>>>> 
>>>> 
>>>> ________________________________
>>>> 
>>>> This message is for the designated recipient only and may contain
>>>> privileged, proprietary, or otherwise confidential information. If you
>>>> have received it in error, please notify the sender immediately and
>>>> delete the original. Any other use of the e-mail by you is prohibited.
>>>> Where allowed by local law, electronic communications with Accenture
>>>> and its affiliates, including e-mail and instant messaging (including
>>>> content), may be scanned by our systems for the purposes of
>>>> information security and assessment of internal compliance with
>>>> Accenture policy.
>>>> ______________________________________________________________________
>>>> _
>>>> ___
>>>> ____________
>>>> 
>>>> www.accenture.com
>>> 
>>> 
>>> 
>>> ________________________________
>>> 
>>> This message is for the designated recipient only and may contain
>>> privileged, proprietary, or otherwise confidential information. If you
>>> have received it in error, please notify the sender immediately and
>>> delete the original. Any other use of the e-mail by you is prohibited.
>>> Where allowed by local law, electronic communications with Accenture
>>> and its affiliates, including e-mail and instant messaging (including
>>> content), may be scanned by our systems for the purposes of information
>>> security and assessment of internal compliance with Accenture policy.
>>> _______________________________________________________________________
>>> ___
>>> ____________
>>> 
>>> www.accenture.com
>> 
>> 
>> 
>> ________________________________
>> 
>> This message is for the designated recipient only and may contain
>> privileged, proprietary, or otherwise confidential information. If you
>> have received it in error, please notify the sender immediately and
>> delete the original. Any other use of the e-mail by you is prohibited.
>> Where allowed by local law, electronic communications with Accenture and
>> its affiliates, including e-mail and instant messaging (including
>> content), may be scanned by our systems for the purposes of information
>> security and assessment of internal compliance with Accenture policy.
>> __________________________________________________________________________
>> ____________
>> 
>> www.accenture.com
> 

Reply via email to