Fwd: XML to TEXT

2014-02-12 Thread Ranjini Rathinam
> > Please help to convert this xml to text. >> >> >> I have the attached the xml. Please find the attachement. >> >> Some student has two address tag and some student has one address tag and >> some student dont have address tag tag. >> >> I need to convert the xml into string. >> >> this is my d

Test hadoop code on the cloud

2014-02-12 Thread Andrea Barbato
Hi! I need to test my hadoop code on a cluster, what is the simplest way to do this on the cloud? Is there any way to do it for free? Thank in advance

Chian Jobs in C++ with Pipes

2014-02-12 Thread Massimo Simoniello
Hi, I'm using Hadoop Pipes and I want to chain two jobs (job1 and job2). Is it possible? I use the FileInputFormat.addInputPath() and FileOutputFormat.setOutputPath() functions to do it in Java, but I want to know if there is some way for do it in C++ with pipes. Thanks in advance, Massimo

Re: Test hadoop code on the cloud

2014-02-12 Thread Zhao Xiaoguang
I think you can test it in Amazon EC2 with pseudo distribute, it support 1 tiny instance for 1 year free. Send From My Macbook On Feb 12, 2014, at 6:29 PM, Andrea Barbato wrote: > Hi! > I need to test my hadoop code on a cluster, > what is the simplest way to do this on the cloud? > Is the

Re: Test hadoop code on the cloud

2014-02-12 Thread Andrea Barbato
Thanks for the answer, but if i want to test my code on a full distributed installation? (for more accurate performance) 2014-02-12 13:01 GMT+01:00 Zhao Xiaoguang : > I think you can test it in Amazon EC2 with pseudo distribute, it support 1 > tiny instance for 1 year free. > > > Send From My Ma

Re: XML to TEXT

2014-02-12 Thread Shekhar Sharma
Which input format you are using . Use xml input format. On 3 Jan 2014 10:47, "Ranjini Rathinam" wrote: > Hi, > > Need to convert XML into text using mapreduce. > > I have used DOM and SAX parser. > > After using SAX Builder in mapper class. the child node act as root > Element. > > While seeing

Re: Test hadoop code on the cloud

2014-02-12 Thread Silvina Caíno Lores
You can check Amazon Elastic MapReduce, which comes preconfigured on EC2 but you need to pay a little por it, or make your custom instalation on EC2 (beware that EC2 instances come with nothing but really basic shell tools on it, so it may take a while to get it running). Amazon's free tier allows

Re: Chian Jobs in C++ with Pipes

2014-02-12 Thread Silvina Caíno Lores
I've been dealing with a similar situation and I haven't found other solution rather than launching two independent jobs (with a script or whatever you like), letting the output of the first be the input of the last. If you find any other option please let me know. Regards On 12 February 2014 12

Re: Test hadoop code on the cloud

2014-02-12 Thread Jay Vyas
As a slightly more advanced option for OpenStack people: Consider trying savanna (Hadoop provisioned on top of open stack) as well. On Wed, Feb 12, 2014 at 10:23 AM, Silvina Caíno Lores wrote: > You can check Amazon Elastic MapReduce, which comes preconfigured on EC2 > but you need to pay a lit

Re: Chian Jobs in C++ with Pipes

2014-02-12 Thread Massimo Simoniello
Yes, of course. It's a solution but I need all jobs in a single file like in java.. Can anyone help me? 2014-02-12 16:34 GMT+01:00 Silvina Caíno Lores : > I've been dealing with a similar situation and I haven't found other > solution rather than launching two independent jobs (with a script or

RE: very long timeout on failed RM connect

2014-02-12 Thread John Lilley
Setting conf.set("yarn.resourcemanager.connect.max-wait.ms", "500"); conf.set("yarn.resourcemanager.connect.retry-interval.ms", "500"); still results in a wait of around 15 seconds. Setting this: conf.set("ipc.client.connect.max.retries", "2"); Also does not help. Is there a retry para

Re: Compression codec com.hadoop.compression.lzo.LzoCodec not found

2014-02-12 Thread Ted Yu
What's the value for "io.compression.codecs" config parameter ? Thanks On Tue, Feb 11, 2014 at 10:11 PM, Li Li wrote: > I am runing example of wordcout but encount an exception: > I googled and know lzo compression's license is incompatible with apache's > so it's not built in. > the question

Re: Can Yarn AppMaster move Container logs to hdfs?

2014-02-12 Thread Jian He
Hi Emmanuel, log aggregation now only aggregates finished apps logs onto hdfs. there's no way as of now to support running apps. that'll be a to-do feature in the feature. Jian On Mon, Feb 10, 2014 at 11:53 AM, Emmanuel Espina wrote: > Sorry when I said log running I meant LONG running, that i

Re: Compression codec com.hadoop.compression.lzo.LzoCodec not found

2014-02-12 Thread Zhijie Shen
For the codecs, you can choose among org.apache.hadoop.io.compress.*Codec. LzoCodec has been moved out of Hadoop (see HADOOP-4874). - Zhijie On Wed, Feb 12, 2014 at 10:54 AM, Ted Yu wrote: > What's the value for "io.compression.codecs" config parameter ? > > Thanks > > > On Tue, Feb 11, 2014 a

Re: very long timeout on failed RM connect

2014-02-12 Thread Jian He
ipc.client.connect.retry.interval sets the underlying ipc retry interval yarn.resourcemanager.connect.retry-interval.ms sets the upper layer clientRMProxy retry interval. Each clientRMProxy retry includes one full round of retries of the underlying ipc. In each ClientRMProxy retry, the max numbe

Re: Compression codec com.hadoop.compression.lzo.LzoCodec not found

2014-02-12 Thread Li Li
io.compression.codecs org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec,org.apache.hadoop.io.compress.BZip2Codec,com.hadoop.compression.lzo.LzoCodec,com.hadoop.compression.lzo.LzopCodec A list of the compression codec classes that can be used

OPENFORWRITE Files issue

2014-02-12 Thread Xiao Li
Say I have a text file on hdfs in "OPENFORWRITE, HEALTHY" status. some process is appending to it. It has 4 lines in it. hadoop fs -cat /file | wc -l 4 However when I do a wordcount on this file, only first line is visible to the mapreduce. Similar in hive when i do "select count(*) fr

Re: Compression codec com.hadoop.compression.lzo.LzoCodec not found

2014-02-12 Thread Ted Yu
Please remove LzoCodec from config. Cheers On Feb 12, 2014, at 5:12 PM, Li Li wrote: > > io.compression.codecs > > org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec,org.apache.hadoop.io.compress.BZip2Codec,com.hadoop.compression.lzo.LzoCodec,com.hadoop.comp

Re: Compression codec com.hadoop.compression.lzo.LzoCodec not found

2014-02-12 Thread Li Li
thanks. it's correct now. On Thu, Feb 13, 2014 at 9:37 AM, Ted Yu wrote: > Please remove LzoCodec from config. > > Cheers > > On Feb 12, 2014, at 5:12 PM, Li Li wrote: > >> >> io.compression.codecs >> >> org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec,org.a

hadoop 2.2.0 QJM exception : NoClassDefFoundError: org/apache/hadoop/hdfs/server/namenode/FSImage

2014-02-12 Thread Henry Hung
Hi All, I don't know why the journal node logs has this weird "NoClassDefFoundError: org/apache/hadoop/hdfs/server/namenode/FSImage" exception. This error occurs each time I switch my namenode from standby to active 2014-02-13 10:34:47,873 INFO org.apache.hadoop.hdfs.server.namenode.FileJournal

(Solved) hadoop 2.2.0 QJM exception : NoClassDefFoundError: org/apache/hadoop/hdfs/server/namenode/FSImage

2014-02-12 Thread Henry Hung
Dear All, Sorry, I found the root cause for this problem, it appears that I overwrite the hadoop-hdfs-2.2.0.jar with my own custom jar, but forgot to restart the journal node process, so the process cannot find the FSImage class, but it actually there inside my custom jar. Note to myself: make

Unable to load native-hadoop library for your platform

2014-02-12 Thread xeon Mailinglist
I am trying to run an example and I get the following error: HadoopMaster-nh:~# /root/Programs/hadoop/bin/hdfs dfs -count /wiki OpenJDK 64-Bit Server VM warning: You have loaded library /root/Programs/hadoop-2.0.5-alpha/lib/native/libhadoop.so.1.0.0 which might have disabled stack guard. The VM wi

RE: Unable to load native-hadoop library for your platform

2014-02-12 Thread Steve Kallestad
Funny, I was just trying to add something to the wiki addressing this. These instructions are for 2.2, but I imagine that 2.0.5 is probably very similar. If the formatting doesn't come through for whatever reason, I posted the same thing here: http://answers.splunk.com/answers/118174/hunk-

Password not found for ApplicationAttempt

2014-02-12 Thread Anfernee Xu
My MR job failed due to below error, I'm running YARN 2.2.0 release. Does anybody know what the error means and how to fix it? 2014-02-12 18:25:31,748 ERROR [main] org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException as:xinx (auth:SIMPLE) cause:org.apache.hadoop.ipc.RemoteE